var/home/core/zuul-output/0000755000175000017500000000000015067454053014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067470007015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005127123515067470000017702 0ustar rootrootOct 02 10:51:56 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 10:51:56 crc restorecon[4670]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:56 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 10:51:57 crc restorecon[4670]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 10:51:58 crc kubenswrapper[4751]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.937094 4751 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951496 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951529 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951546 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951554 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951563 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951575 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951586 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951596 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951605 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951614 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951623 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951631 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951639 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951647 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951662 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951670 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951678 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951686 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951694 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951702 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951710 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951717 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951725 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951734 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951741 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951749 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951757 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951774 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951781 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951791 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951799 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951810 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951818 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951826 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951834 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951842 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951850 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951857 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951874 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951884 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951894 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951905 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951914 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951923 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951933 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951942 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951951 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951962 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951972 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951984 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.951994 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952012 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952028 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952043 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952056 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952070 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952082 4751 feature_gate.go:330] unrecognized feature gate: Example Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952095 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952107 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952118 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952135 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952148 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952208 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952223 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952235 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952252 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952265 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952277 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952288 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952300 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.952312 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954293 4751 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954355 4751 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954376 4751 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954388 4751 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954402 4751 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954417 4751 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954441 4751 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954467 4751 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954481 4751 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954493 4751 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954508 4751 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954521 4751 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954533 4751 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954545 4751 flags.go:64] FLAG: --cgroup-root="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954556 4751 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954568 4751 flags.go:64] FLAG: --client-ca-file="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954579 4751 flags.go:64] FLAG: --cloud-config="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954591 4751 flags.go:64] FLAG: --cloud-provider="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954602 4751 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954623 4751 flags.go:64] FLAG: --cluster-domain="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954633 4751 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954647 4751 flags.go:64] FLAG: --config-dir="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954659 4751 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954672 4751 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954688 4751 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954699 4751 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954711 4751 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954724 4751 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954736 4751 flags.go:64] FLAG: --contention-profiling="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954749 4751 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954761 4751 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954774 4751 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954786 4751 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954802 4751 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954814 4751 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954825 4751 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954837 4751 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954849 4751 flags.go:64] FLAG: --enable-server="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954860 4751 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954875 4751 flags.go:64] FLAG: --event-burst="100" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954887 4751 flags.go:64] FLAG: --event-qps="50" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954899 4751 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954912 4751 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954924 4751 flags.go:64] FLAG: --eviction-hard="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954938 4751 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954950 4751 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954961 4751 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954973 4751 flags.go:64] FLAG: --eviction-soft="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954985 4751 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.954996 4751 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955008 4751 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955020 4751 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955032 4751 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955043 4751 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955054 4751 flags.go:64] FLAG: --feature-gates="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955070 4751 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955082 4751 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955094 4751 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955104 4751 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955115 4751 flags.go:64] FLAG: --healthz-port="10248" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955125 4751 flags.go:64] FLAG: --help="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955134 4751 flags.go:64] FLAG: --hostname-override="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955144 4751 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955155 4751 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955219 4751 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955237 4751 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955246 4751 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955256 4751 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955266 4751 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955275 4751 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955284 4751 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955294 4751 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955304 4751 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955313 4751 flags.go:64] FLAG: --kube-reserved="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955323 4751 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955332 4751 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955341 4751 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955350 4751 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955360 4751 flags.go:64] FLAG: --lock-file="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955369 4751 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955378 4751 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955389 4751 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955406 4751 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955417 4751 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955427 4751 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955436 4751 flags.go:64] FLAG: --logging-format="text" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955445 4751 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955455 4751 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955464 4751 flags.go:64] FLAG: --manifest-url="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955475 4751 flags.go:64] FLAG: --manifest-url-header="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955503 4751 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955514 4751 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955526 4751 flags.go:64] FLAG: --max-pods="110" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955535 4751 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955545 4751 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955555 4751 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955564 4751 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955574 4751 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955583 4751 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955594 4751 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955621 4751 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955630 4751 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955640 4751 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955649 4751 flags.go:64] FLAG: --pod-cidr="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955660 4751 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955674 4751 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955683 4751 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955693 4751 flags.go:64] FLAG: --pods-per-core="0" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955702 4751 flags.go:64] FLAG: --port="10250" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955713 4751 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955722 4751 flags.go:64] FLAG: --provider-id="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955732 4751 flags.go:64] FLAG: --qos-reserved="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955741 4751 flags.go:64] FLAG: --read-only-port="10255" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955750 4751 flags.go:64] FLAG: --register-node="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955760 4751 flags.go:64] FLAG: --register-schedulable="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955769 4751 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955786 4751 flags.go:64] FLAG: --registry-burst="10" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955795 4751 flags.go:64] FLAG: --registry-qps="5" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955804 4751 flags.go:64] FLAG: --reserved-cpus="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955814 4751 flags.go:64] FLAG: --reserved-memory="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955826 4751 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955836 4751 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955845 4751 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955857 4751 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955869 4751 flags.go:64] FLAG: --runonce="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955887 4751 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955909 4751 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955922 4751 flags.go:64] FLAG: --seccomp-default="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955933 4751 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955947 4751 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955960 4751 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955972 4751 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955984 4751 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.955995 4751 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956006 4751 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956018 4751 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956051 4751 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956063 4751 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956075 4751 flags.go:64] FLAG: --system-cgroups="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956086 4751 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956107 4751 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956118 4751 flags.go:64] FLAG: --tls-cert-file="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956129 4751 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956149 4751 flags.go:64] FLAG: --tls-min-version="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956161 4751 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956241 4751 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956257 4751 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956269 4751 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956281 4751 flags.go:64] FLAG: --v="2" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956298 4751 flags.go:64] FLAG: --version="false" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956312 4751 flags.go:64] FLAG: --vmodule="" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956326 4751 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.956339 4751 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956739 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956794 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956805 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956814 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956822 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956830 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956838 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956845 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956854 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956862 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956873 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956882 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956890 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956899 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956906 4751 feature_gate.go:330] unrecognized feature gate: Example Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956915 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956922 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956930 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956941 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956952 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956960 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956968 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956976 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956986 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.956994 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957003 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957011 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957019 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957028 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957038 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957046 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957056 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957065 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957073 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957082 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957090 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957098 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957105 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957115 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957123 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957132 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957139 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957147 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957155 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957163 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957172 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957217 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957225 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957233 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957241 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957249 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957256 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957264 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957271 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957282 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957292 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957300 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957309 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957317 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957325 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957333 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957341 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957349 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957357 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957364 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957372 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957379 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957387 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957407 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957415 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.957424 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.958423 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.971310 4751 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.971353 4751 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971473 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971486 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971495 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971505 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971515 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971524 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971533 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971541 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971550 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971558 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971566 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971575 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971585 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971596 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971606 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971615 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971623 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971631 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971639 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971649 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971657 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971664 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971672 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971680 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971689 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971696 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971704 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971712 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971721 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971730 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971738 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971748 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971758 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971766 4751 feature_gate.go:330] unrecognized feature gate: Example Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971776 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971785 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971792 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971800 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971808 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971816 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971825 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971833 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971840 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971848 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971856 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971864 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971871 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971879 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971887 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971894 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971902 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971912 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971924 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971932 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971939 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971948 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971956 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971963 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971971 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971978 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971987 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.971995 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972002 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972010 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972017 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972025 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972033 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972043 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972052 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972062 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972071 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.972085 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972395 4751 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972410 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972422 4751 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972432 4751 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972441 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972449 4751 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972458 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972466 4751 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972473 4751 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972481 4751 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972489 4751 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972497 4751 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972507 4751 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972516 4751 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972525 4751 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972533 4751 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972541 4751 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972550 4751 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972558 4751 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972567 4751 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972576 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972583 4751 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972591 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972598 4751 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972606 4751 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972614 4751 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972621 4751 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972628 4751 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972636 4751 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972644 4751 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972651 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972659 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972666 4751 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972674 4751 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972684 4751 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972692 4751 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972699 4751 feature_gate.go:330] unrecognized feature gate: Example Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972707 4751 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972714 4751 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972722 4751 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972730 4751 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972738 4751 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972745 4751 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972754 4751 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972763 4751 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972772 4751 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972780 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972788 4751 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972795 4751 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972803 4751 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972811 4751 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972819 4751 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972827 4751 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972835 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972843 4751 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972851 4751 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972859 4751 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972866 4751 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972874 4751 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972881 4751 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972888 4751 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972896 4751 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972904 4751 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972911 4751 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972921 4751 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972950 4751 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972962 4751 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972971 4751 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972980 4751 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972988 4751 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 10:51:58 crc kubenswrapper[4751]: W1002 10:51:58.972997 4751 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.973009 4751 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 10:51:58 crc kubenswrapper[4751]: I1002 10:51:58.973295 4751 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.003998 4751 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.004218 4751 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.006135 4751 server.go:997] "Starting client certificate rotation" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.006214 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.006498 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-05 06:42:35.532112059 +0000 UTC Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.006677 4751 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2275h50m36.525441061s for next certificate rotation Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.086075 4751 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.089136 4751 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.145578 4751 log.go:25] "Validated CRI v1 runtime API" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.234154 4751 log.go:25] "Validated CRI v1 image API" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.241212 4751 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.251329 4751 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-10-45-59-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.251373 4751 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.271704 4751 manager.go:217] Machine: {Timestamp:2025-10-02 10:51:59.269536369 +0000 UTC m=+1.323762859 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da BootID:9439a9ad-297a-4aaf-997f-74421bdf3adb Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:cf:35:42 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:cf:35:42 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:09:65:45 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:fe:ad:6a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:12:cb:f1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e2:b6:85 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:72:8c:24:37:4f:c7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ca:c6:94:76:2c:4e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.272069 4751 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.272412 4751 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.273150 4751 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.273405 4751 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.273448 4751 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.273721 4751 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.273735 4751 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.274316 4751 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.274357 4751 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.274573 4751 state_mem.go:36] "Initialized new in-memory state store" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.274678 4751 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.315879 4751 kubelet.go:418] "Attempting to sync node with API server" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.315932 4751 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.315983 4751 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.316008 4751 kubelet.go:324] "Adding apiserver pod source" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.316036 4751 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.330451 4751 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.333275 4751 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 10:51:59 crc kubenswrapper[4751]: W1002 10:51:59.334126 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:51:59 crc kubenswrapper[4751]: W1002 10:51:59.334160 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.334307 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.334320 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.368221 4751 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.371849 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372021 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372132 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372286 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372408 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372543 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372650 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372767 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372876 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.372978 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.373101 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.373254 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.374394 4751 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.375440 4751 server.go:1280] "Started kubelet" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.376036 4751 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.376198 4751 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.377702 4751 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.378125 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:51:59 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380047 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380154 4751 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380478 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 05:03:33.521645014 +0000 UTC Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380560 4751 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1194h11m34.141091768s for next certificate rotation Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.380767 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380903 4751 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.380944 4751 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.381130 4751 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.381646 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="200ms" Oct 02 10:51:59 crc kubenswrapper[4751]: W1002 10:51:59.381912 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.382022 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.382165 4751 factory.go:55] Registering systemd factory Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.382250 4751 factory.go:221] Registration of the systemd container factory successfully Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.389609 4751 factory.go:153] Registering CRI-O factory Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.389722 4751 factory.go:221] Registration of the crio container factory successfully Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.393522 4751 server.go:460] "Adding debug handlers to kubelet server" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.394600 4751 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.394668 4751 factory.go:103] Registering Raw factory Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.394860 4751 manager.go:1196] Started watching for new ooms in manager Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.417085 4751 manager.go:319] Starting recovery of all containers Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429291 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429850 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429874 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429895 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429915 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429936 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429955 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429977 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.429999 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430018 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430036 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430057 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430076 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430136 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430154 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430205 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430228 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430247 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430265 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430283 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430307 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430327 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430349 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430369 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430391 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430457 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430481 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430502 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430524 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430545 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430620 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430641 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430665 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430685 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430706 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430726 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430745 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430774 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430796 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430815 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430834 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430854 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430872 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430891 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430910 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.430988 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431011 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431033 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431055 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431078 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431098 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431120 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431146 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431173 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431328 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431351 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431373 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431392 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431411 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431430 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431447 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431465 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431482 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431501 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431522 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431542 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431562 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431580 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431597 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431616 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431634 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431657 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431680 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431722 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431742 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431760 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431779 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431809 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431832 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431860 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431891 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431916 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431936 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431955 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431974 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.431994 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432015 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432034 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432053 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432072 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432091 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432110 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432128 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432147 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432166 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432397 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432422 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432440 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432460 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432478 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432497 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432516 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432535 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432554 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432582 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432603 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432625 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432646 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432665 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432686 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432707 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432726 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432747 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432766 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432786 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432805 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432823 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432842 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432861 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432881 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432900 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432919 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432939 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432961 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432979 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.432999 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433104 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433123 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433144 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433162 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433211 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433231 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433250 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433271 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433292 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433313 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433331 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433355 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433373 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433391 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433409 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433427 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433446 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433465 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433487 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433506 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433526 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433545 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433564 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433586 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433608 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433626 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433645 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433663 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433683 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433706 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433723 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433743 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433768 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433813 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433846 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433865 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433883 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433902 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433921 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433940 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433959 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433979 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.433999 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.434023 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.434044 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.434063 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.434085 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.448158 4751 manager.go:324] Recovery completed Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.417361 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.163:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa716775e1345 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 10:51:59.375373125 +0000 UTC m=+1.429599615,LastTimestamp:2025-10-02 10:51:59.375373125 +0000 UTC m=+1.429599615,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462148 4751 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462266 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462294 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462320 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462340 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462362 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462383 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462404 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462427 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462445 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462502 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462524 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462545 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462571 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462591 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462612 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462633 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462654 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462675 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462695 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462714 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462732 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462750 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462769 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462838 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462888 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462915 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462938 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462964 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.462990 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463015 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463043 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463063 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463083 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463105 4751 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463151 4751 reconstruct.go:97] "Volume reconstruction finished" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463165 4751 reconciler.go:26] "Reconciler: start to sync state" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.463689 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.465381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.465452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.465466 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.467006 4751 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.467028 4751 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.467086 4751 state_mem.go:36] "Initialized new in-memory state store" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.481930 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.545219 4751 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.548164 4751 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.548776 4751 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.548826 4751 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.548880 4751 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 10:51:59 crc kubenswrapper[4751]: W1002 10:51:59.551177 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.551374 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.582303 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="400ms" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.582404 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.649851 4751 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.683107 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.749489 4751 policy_none.go:49] "None policy: Start" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.751249 4751 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.751331 4751 state_mem.go:35] "Initializing new in-memory state store" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.783665 4751 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.850673 4751 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.874964 4751 manager.go:334] "Starting Device Plugin manager" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.875210 4751 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.875248 4751 server.go:79] "Starting device plugin registration server" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.875869 4751 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.875905 4751 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.876273 4751 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.876436 4751 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.876450 4751 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.888281 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.976975 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.978417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.978470 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.978489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:51:59 crc kubenswrapper[4751]: I1002 10:51:59.978526 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.979290 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:51:59 crc kubenswrapper[4751]: E1002 10:51:59.983117 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="800ms" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.180427 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.182764 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.182822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.182839 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.182874 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.183615 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.216414 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.216521 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.250944 4751 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.251088 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.252625 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.252675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.252693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.252880 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.253295 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.253360 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254442 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254505 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254529 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254664 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254706 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254759 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254807 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.254857 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256120 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256216 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256363 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256463 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256510 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256903 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.256997 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257826 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257842 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.257987 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.258335 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.258542 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.258631 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260447 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260496 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260514 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260854 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260940 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.260963 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.261347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.261429 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.262835 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.262917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.262943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.335551 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.335692 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377055 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377123 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377157 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377250 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377306 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377344 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377507 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377653 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377689 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377722 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377805 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377866 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377905 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.377933 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.378986 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.472630 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.472775 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479619 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479701 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479734 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479765 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479808 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479839 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479868 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479868 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479903 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479928 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479942 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479986 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479865 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480042 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480095 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.479980 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480017 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480145 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480020 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480053 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480261 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480259 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480359 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480498 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480579 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480895 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.480973 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.481008 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.481140 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.481281 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.584286 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.586200 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.586282 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.586302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.586343 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.586794 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.588006 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.610159 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.632379 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.648328 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: I1002 10:52:00.659716 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.676603 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-067e680971cc37e60df9aae5b6e6eb0f78cf3c351aec42d9c380afd2fd110dde WatchSource:0}: Error finding container 067e680971cc37e60df9aae5b6e6eb0f78cf3c351aec42d9c380afd2fd110dde: Status 404 returned error can't find the container with id 067e680971cc37e60df9aae5b6e6eb0f78cf3c351aec42d9c380afd2fd110dde Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.679516 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-98a4cf70a694f21fc721fa3e3a65d83e12e8200ecde6514358cc8368c4e4d29d WatchSource:0}: Error finding container 98a4cf70a694f21fc721fa3e3a65d83e12e8200ecde6514358cc8368c4e4d29d: Status 404 returned error can't find the container with id 98a4cf70a694f21fc721fa3e3a65d83e12e8200ecde6514358cc8368c4e4d29d Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.693546 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d0fd931e615692282fc15fcf975b30ea5ce0f1d4e017b83fef7db54728e0aadc WatchSource:0}: Error finding container d0fd931e615692282fc15fcf975b30ea5ce0f1d4e017b83fef7db54728e0aadc: Status 404 returned error can't find the container with id d0fd931e615692282fc15fcf975b30ea5ce0f1d4e017b83fef7db54728e0aadc Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.694833 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f1a99459f1dd6e05d22a00e11890238dba44925273cfb0cb87456b8a9a242dbc WatchSource:0}: Error finding container f1a99459f1dd6e05d22a00e11890238dba44925273cfb0cb87456b8a9a242dbc: Status 404 returned error can't find the container with id f1a99459f1dd6e05d22a00e11890238dba44925273cfb0cb87456b8a9a242dbc Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.700706 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6be5ecb90b14e05b1ce68001437de169e3c4f77d2dddff8c4086151e17af085c WatchSource:0}: Error finding container 6be5ecb90b14e05b1ce68001437de169e3c4f77d2dddff8c4086151e17af085c: Status 404 returned error can't find the container with id 6be5ecb90b14e05b1ce68001437de169e3c4f77d2dddff8c4086151e17af085c Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.784568 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="1.6s" Oct 02 10:52:00 crc kubenswrapper[4751]: W1002 10:52:00.827652 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:00 crc kubenswrapper[4751]: E1002 10:52:00.827781 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.379452 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.387835 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.389730 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.389768 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.389788 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.389813 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:01 crc kubenswrapper[4751]: E1002 10:52:01.390279 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.557672 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6be5ecb90b14e05b1ce68001437de169e3c4f77d2dddff8c4086151e17af085c"} Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.562699 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f1a99459f1dd6e05d22a00e11890238dba44925273cfb0cb87456b8a9a242dbc"} Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.563838 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d0fd931e615692282fc15fcf975b30ea5ce0f1d4e017b83fef7db54728e0aadc"} Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.564998 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"067e680971cc37e60df9aae5b6e6eb0f78cf3c351aec42d9c380afd2fd110dde"} Oct 02 10:52:01 crc kubenswrapper[4751]: I1002 10:52:01.566287 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"98a4cf70a694f21fc721fa3e3a65d83e12e8200ecde6514358cc8368c4e4d29d"} Oct 02 10:52:02 crc kubenswrapper[4751]: W1002 10:52:02.033633 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:02 crc kubenswrapper[4751]: E1002 10:52:02.033747 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.379398 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:02 crc kubenswrapper[4751]: E1002 10:52:02.386134 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="3.2s" Oct 02 10:52:02 crc kubenswrapper[4751]: W1002 10:52:02.391149 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:02 crc kubenswrapper[4751]: E1002 10:52:02.391293 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:02 crc kubenswrapper[4751]: W1002 10:52:02.521402 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:02 crc kubenswrapper[4751]: E1002 10:52:02.521477 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.991314 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.993516 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.993598 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.993617 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:02 crc kubenswrapper[4751]: I1002 10:52:02.993661 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:02 crc kubenswrapper[4751]: E1002 10:52:02.994558 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:52:03 crc kubenswrapper[4751]: W1002 10:52:03.318425 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:03 crc kubenswrapper[4751]: E1002 10:52:03.318564 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.380319 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.576349 4751 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55" exitCode=0 Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.576456 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.576519 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.578155 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.578213 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.578227 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.579935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.579991 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.581828 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5" exitCode=0 Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.581905 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.581989 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.583310 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.583369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.583388 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.584585 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c" exitCode=0 Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.584807 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.584940 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.586510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.586604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.586629 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.589091 4751 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d" exitCode=0 Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.589148 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d"} Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.589442 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.589516 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591703 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591741 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:03 crc kubenswrapper[4751]: I1002 10:52:03.591757 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.379401 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.596538 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.596605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.596635 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.597948 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.597999 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.598013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.598827 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.600826 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803" exitCode=0 Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.600887 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.600980 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.601874 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.601919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.601932 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.603441 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.603538 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.604440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.604486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.604507 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.605858 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.605906 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2"} Oct 02 10:52:04 crc kubenswrapper[4751]: I1002 10:52:04.977993 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.076388 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.291660 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.292298 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.292384 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.378996 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:05 crc kubenswrapper[4751]: E1002 10:52:05.587422 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.163:6443: connect: connection refused" interval="6.4s" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.613231 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6"} Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.613287 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f"} Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.616821 4751 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c" exitCode=0 Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.616925 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c"} Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.616985 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.618350 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.618401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.618418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.620487 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.620745 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec"} Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.620881 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.620890 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621667 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621729 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621940 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.621991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.622494 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.622542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:05 crc kubenswrapper[4751]: I1002 10:52:05.622560 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.195588 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.197583 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.197644 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.197669 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.197710 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:06 crc kubenswrapper[4751]: E1002 10:52:06.198438 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.163:6443: connect: connection refused" node="crc" Oct 02 10:52:06 crc kubenswrapper[4751]: W1002 10:52:06.220612 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:06 crc kubenswrapper[4751]: E1002 10:52:06.220728 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.379372 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.626579 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f"} Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.626664 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.626662 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.626608 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627576 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627730 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627746 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:06 crc kubenswrapper[4751]: I1002 10:52:06.627754 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:06 crc kubenswrapper[4751]: W1002 10:52:06.863843 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:06 crc kubenswrapper[4751]: E1002 10:52:06.863970 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.379094 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.632771 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f"} Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.632924 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.634233 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.634259 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.634267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.637718 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"053c003e68ab876f5ce044e94c140083a0387c5e2fb40f70eb57b96551b28cec"} Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.637792 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e2fce79d0c287c78276c423a4da66591c96c894ab527adcb72fc07bd2d1e0cc1"} Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.637813 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2a29f9a94eda077e73d5b97cc2454a48ec497825258a0f78f5f4cddf7179fe23"} Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.637831 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.639050 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.639096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:07 crc kubenswrapper[4751]: I1002 10:52:07.639106 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:07 crc kubenswrapper[4751]: W1002 10:52:07.753076 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:07 crc kubenswrapper[4751]: E1002 10:52:07.753258 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:08 crc kubenswrapper[4751]: E1002 10:52:08.032256 4751 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.163:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa716775e1345 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 10:51:59.375373125 +0000 UTC m=+1.429599615,LastTimestamp:2025-10-02 10:51:59.375373125 +0000 UTC m=+1.429599615,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 10:52:08 crc kubenswrapper[4751]: W1002 10:52:08.080019 4751 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:08 crc kubenswrapper[4751]: E1002 10:52:08.080120 4751 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.163:6443: connect: connection refused" logger="UnhandledError" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.379336 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.163:6443: connect: connection refused Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.565441 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.566004 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.566089 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.646833 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ede217fd4c602d51b6800d65fb9132b04b8ad8c0bc0e051016bae140e015321b"} Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.646895 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5bac1f5f3060b21b391cae56b30d714b25dd6e311c123523acaf77901d8e1354"} Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.646975 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.648384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.648426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.648443 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.648839 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.651478 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f" exitCode=255 Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.651525 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f"} Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.651572 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.652536 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.652579 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.652597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:08 crc kubenswrapper[4751]: I1002 10:52:08.653259 4751 scope.go:117] "RemoveContainer" containerID="878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.657462 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.661242 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.661300 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6"} Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.661355 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.661242 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.663853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.663901 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.663920 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.663985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.664017 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:09 crc kubenswrapper[4751]: I1002 10:52:09.664033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:09 crc kubenswrapper[4751]: E1002 10:52:09.888429 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.633967 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.663411 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.663470 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.663590 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.664917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.664976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.664978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.665023 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.665042 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.665004 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.973355 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.973567 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.975201 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.975239 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:10 crc kubenswrapper[4751]: I1002 10:52:10.975258 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:11 crc kubenswrapper[4751]: I1002 10:52:11.665590 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:11 crc kubenswrapper[4751]: I1002 10:52:11.666981 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:11 crc kubenswrapper[4751]: I1002 10:52:11.667035 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:11 crc kubenswrapper[4751]: I1002 10:52:11.667051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:11 crc kubenswrapper[4751]: I1002 10:52:11.936916 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.095949 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.096410 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.098389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.098425 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.098439 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.599392 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.601152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.601227 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.601241 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.601274 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.668564 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.669957 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.669996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:12 crc kubenswrapper[4751]: I1002 10:52:12.670005 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:13 crc kubenswrapper[4751]: I1002 10:52:13.974216 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 10:52:13 crc kubenswrapper[4751]: I1002 10:52:13.974336 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.299532 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.299710 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.301000 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.301039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.301055 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.304476 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.677872 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.679267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.679315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:15 crc kubenswrapper[4751]: I1002 10:52:15.679328 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:19 crc kubenswrapper[4751]: I1002 10:52:19.379225 4751 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 10:52:19 crc kubenswrapper[4751]: I1002 10:52:19.611702 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 10:52:19 crc kubenswrapper[4751]: I1002 10:52:19.611767 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 10:52:19 crc kubenswrapper[4751]: I1002 10:52:19.617353 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 10:52:19 crc kubenswrapper[4751]: I1002 10:52:19.617427 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 10:52:19 crc kubenswrapper[4751]: E1002 10:52:19.888776 4751 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.138527 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.138833 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.140377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.140448 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.140466 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.155894 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.697777 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.699236 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.699300 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:22 crc kubenswrapper[4751]: I1002 10:52:22.699319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.572917 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.573230 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.574859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.574923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.574956 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.580796 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.701153 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.701512 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.702878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.702937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.702955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.973672 4751 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 10:52:23 crc kubenswrapper[4751]: I1002 10:52:23.973764 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 10:52:24 crc kubenswrapper[4751]: E1002 10:52:24.592458 4751 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.594762 4751 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.597345 4751 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.598212 4751 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.598245 4751 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 10:52:24 crc kubenswrapper[4751]: E1002 10:52:24.599434 4751 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.602469 4751 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.614078 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49038->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.614821 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49038->192.168.126.11:17697: read: connection reset by peer" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.615477 4751 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.615543 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.705217 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.705984 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.707784 4751 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6" exitCode=255 Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.707822 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6"} Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.707868 4751 scope.go:117] "RemoveContainer" containerID="878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f" Oct 02 10:52:24 crc kubenswrapper[4751]: I1002 10:52:24.730281 4751 scope.go:117] "RemoveContainer" containerID="d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6" Oct 02 10:52:24 crc kubenswrapper[4751]: E1002 10:52:24.730516 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.333010 4751 apiserver.go:52] "Watching apiserver" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.336487 4751 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.336912 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.337766 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.337931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.338160 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.338266 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.338297 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.338454 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.338507 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.338583 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.339993 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.345524 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.345573 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.346265 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.346291 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.346531 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.346944 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.347307 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.347416 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.347803 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.382157 4751 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.386603 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:08Z\\\",\\\"message\\\":\\\"W1002 10:52:07.833951 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 10:52:07.834475 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759402327 cert, and key in /tmp/serving-cert-2091720293/serving-signer.crt, /tmp/serving-cert-2091720293/serving-signer.key\\\\nI1002 10:52:08.202913 1 observer_polling.go:159] Starting file observer\\\\nW1002 10:52:08.207103 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 10:52:08.207270 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:08.207807 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2091720293/tls.crt::/tmp/serving-cert-2091720293/tls.key\\\\\\\"\\\\nF1002 10:52:08.443016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403422 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403528 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403600 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403973 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403967 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.403631 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404065 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404096 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404057 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404122 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404146 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404191 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404212 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404234 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404257 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404285 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404394 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404508 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404895 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.404943 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405208 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405225 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405264 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405291 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405317 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405364 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405390 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405416 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405440 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405462 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405486 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405508 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405529 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405559 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405580 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405600 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405620 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405641 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405661 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405681 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405682 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405702 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405808 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405870 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405924 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406014 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406073 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406129 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406215 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406276 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406328 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406383 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406444 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406494 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406547 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406597 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406650 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406702 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406751 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406801 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407038 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407087 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407135 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407232 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407285 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407338 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407390 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407438 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407461 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407523 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407574 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407626 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407672 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407723 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407774 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407825 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407880 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407930 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407979 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408030 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408078 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408131 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408710 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408776 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408831 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408887 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408935 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409113 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409214 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409272 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409322 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409376 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409430 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409485 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409540 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409592 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409643 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409707 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409755 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409810 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409860 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409912 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409991 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410045 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410095 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410153 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410257 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410312 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410362 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410412 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410510 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410566 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410617 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410667 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410716 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410773 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410827 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410884 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410934 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410985 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411034 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405789 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405825 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411088 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411138 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411229 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411284 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411336 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411394 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411447 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411498 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411546 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411599 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411656 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411712 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411765 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411817 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411870 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411927 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411980 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412034 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412106 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412159 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412248 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412309 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412361 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412520 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412576 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412633 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412691 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412742 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412796 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412855 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412904 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412956 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413008 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413068 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413132 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413220 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413275 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413338 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413393 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413449 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413501 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413552 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413613 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413750 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413813 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413873 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413929 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414022 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414076 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414138 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414259 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414317 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.405988 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406210 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406302 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406407 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406453 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.406631 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407270 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407650 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408043 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408090 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.407808 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408545 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408604 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408588 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408916 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414665 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.408999 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409278 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409331 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409368 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.409432 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410078 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410072 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410193 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410238 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410385 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410409 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410431 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410572 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410618 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410638 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410854 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.410926 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411137 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411120 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411603 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411853 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.411905 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412059 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412235 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412292 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412322 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412605 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412780 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412804 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.412927 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413001 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413243 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413443 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413658 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413678 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413758 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.413928 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414100 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414118 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414280 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414574 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414786 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.415443 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.416040 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.416841 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.417219 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.417268 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.417783 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.418124 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.418659 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.418840 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419216 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419381 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419468 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419506 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419546 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.419511 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.420061 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.420574 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.420817 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.420844 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421036 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421060 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421612 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421619 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421737 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.421919 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422233 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422356 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422603 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422769 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422952 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.422978 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423057 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423483 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.414371 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423604 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423807 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423855 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423894 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.423976 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424017 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424049 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424085 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424092 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424125 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424196 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424252 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424288 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424327 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424520 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424571 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.424597 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425109 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425354 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425406 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425481 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425522 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425560 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425594 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425627 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425659 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425684 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425709 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425737 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425769 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425799 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425827 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425900 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425926 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425949 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425952 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425999 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426022 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426051 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426115 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426151 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426218 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426285 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426311 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426334 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426357 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426414 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426449 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426510 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426543 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426657 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426677 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426694 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426712 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426728 4751 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426745 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426761 4751 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426779 4751 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426795 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426811 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426828 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426844 4751 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426860 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426889 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426902 4751 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426915 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426929 4751 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426942 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426954 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426965 4751 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426978 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426991 4751 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427005 4751 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427020 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427037 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427053 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427070 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427088 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427108 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427127 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427144 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427157 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427194 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427212 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427229 4751 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427242 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427255 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427267 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427281 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427293 4751 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427305 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427316 4751 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427328 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427340 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427352 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427366 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427378 4751 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427390 4751 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427403 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427415 4751 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427426 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427438 4751 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427454 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427466 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427479 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427491 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427504 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427517 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427529 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427542 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427553 4751 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427565 4751 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427577 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427589 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427604 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427618 4751 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427630 4751 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427641 4751 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427653 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427667 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427684 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427701 4751 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427716 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427733 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427748 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427762 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427778 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427796 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427812 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427831 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427849 4751 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427866 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427883 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427900 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427920 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427936 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427952 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427968 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427984 4751 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428000 4751 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428018 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428034 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428050 4751 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428067 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428084 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428142 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428160 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428198 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428236 4751 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428257 4751 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428278 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428296 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428311 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428329 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428345 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428362 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428379 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428397 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428414 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428432 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428451 4751 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429647 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429629 4751 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426382 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426445 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425401 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.426885 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.425916 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429731 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427299 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427336 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427529 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427544 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427701 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.427994 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428094 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428439 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428610 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428649 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.428731 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429090 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429366 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429227 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.429752 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.429904 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.430061 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:25.930033656 +0000 UTC m=+27.984260106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430126 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430347 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430485 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430507 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.430790 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430798 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.430863 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:25.930841638 +0000 UTC m=+27.985068108 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.430939 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.431101 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.431584 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.431820 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.431856 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.431972 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.432011 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.432291 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:25.932270376 +0000 UTC m=+27.986496836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.432347 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.432526 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.432549 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.432565 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.433056 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.433060 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.434333 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.434825 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.437606 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.437844 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.438057 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.438366 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.439580 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.440340 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.440590 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.441233 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.441605 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.442701 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.443797 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.443972 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.444432 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.444576 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.445021 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.445323 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.445689 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.446459 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.447137 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.451030 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.452853 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.453351 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.454279 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.454369 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.454475 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.454657 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.454926 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455160 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455493 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455655 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455713 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455850 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455883 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.455902 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.455909 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456109 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456124 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456138 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456211 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:25.956194204 +0000 UTC m=+28.010420654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.456475 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.456660 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456827 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.456916 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.459414 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.459538 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:25.959518102 +0000 UTC m=+28.013744562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.459893 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.460189 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.461534 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.465766 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.472021 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://878be983243f4629be948698216ae6790eae556975851f5db77b67680f088d9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:08Z\\\",\\\"message\\\":\\\"W1002 10:52:07.833951 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 10:52:07.834475 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759402327 cert, and key in /tmp/serving-cert-2091720293/serving-signer.crt, /tmp/serving-cert-2091720293/serving-signer.key\\\\nI1002 10:52:08.202913 1 observer_polling.go:159] Starting file observer\\\\nW1002 10:52:08.207103 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 10:52:08.207270 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:08.207807 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2091720293/tls.crt::/tmp/serving-cert-2091720293/tls.key\\\\\\\"\\\\nF1002 10:52:08.443016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.475476 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.479358 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.482664 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.485624 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.494846 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.495286 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.504151 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529581 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529638 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529709 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529725 4751 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529803 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529857 4751 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529959 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529978 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529989 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.529998 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530007 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530016 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530024 4751 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530033 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530042 4751 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530072 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530082 4751 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530091 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530100 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530139 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530185 4751 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530197 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530208 4751 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530217 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530242 4751 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530252 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530262 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530293 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530312 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530362 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530376 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530388 4751 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530400 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530412 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530424 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530437 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530448 4751 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530460 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530472 4751 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530484 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530496 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530510 4751 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530522 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530535 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530548 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530562 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530574 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530586 4751 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530597 4751 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530609 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530622 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530635 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530647 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530660 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530672 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530684 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530697 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530709 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530721 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530733 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530745 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530757 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530768 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530780 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530793 4751 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530807 4751 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530819 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530830 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530842 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530855 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530867 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530880 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530893 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530907 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530920 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530934 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530947 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530959 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530971 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530983 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.530995 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531007 4751 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531019 4751 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531030 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531043 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531057 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.531070 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.554786 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.555540 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.557681 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.558768 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.560214 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.561647 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.563036 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.565112 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.566710 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.568891 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.569981 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.572684 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.573948 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.575305 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.577495 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.578749 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.580985 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.581826 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.583870 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.585870 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.586553 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.587914 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.588557 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.590002 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.590631 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.591487 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.592974 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.593697 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.595075 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.595725 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.597126 4751 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.597386 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.601078 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.603109 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.604001 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.607461 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.609925 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.611704 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.612394 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.613514 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.614030 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.615273 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.615881 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.616864 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.617374 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.618460 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.619015 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.620372 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.620942 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.621934 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.622504 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.623482 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.624026 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.624527 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.665831 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.681763 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 10:52:25 crc kubenswrapper[4751]: W1002 10:52:25.682344 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ba19099759615be34cb736d0ef45c7d847a354d3f3309c56332931486d071c4c WatchSource:0}: Error finding container ba19099759615be34cb736d0ef45c7d847a354d3f3309c56332931486d071c4c: Status 404 returned error can't find the container with id ba19099759615be34cb736d0ef45c7d847a354d3f3309c56332931486d071c4c Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.691304 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 10:52:25 crc kubenswrapper[4751]: W1002 10:52:25.692691 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-54ece9f21b5b34a9e375db2097a383196c6cfc3973f78c66b7e42391659ae055 WatchSource:0}: Error finding container 54ece9f21b5b34a9e375db2097a383196c6cfc3973f78c66b7e42391659ae055: Status 404 returned error can't find the container with id 54ece9f21b5b34a9e375db2097a383196c6cfc3973f78c66b7e42391659ae055 Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.712504 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"54ece9f21b5b34a9e375db2097a383196c6cfc3973f78c66b7e42391659ae055"} Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.715216 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ba19099759615be34cb736d0ef45c7d847a354d3f3309c56332931486d071c4c"} Oct 02 10:52:25 crc kubenswrapper[4751]: W1002 10:52:25.715420 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-8539cc47cb7cefa10113c2bec7ae704568f0c0e1b2342982ae71b861f05d2c3c WatchSource:0}: Error finding container 8539cc47cb7cefa10113c2bec7ae704568f0c0e1b2342982ae71b861f05d2c3c: Status 404 returned error can't find the container with id 8539cc47cb7cefa10113c2bec7ae704568f0c0e1b2342982ae71b861f05d2c3c Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.718131 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.721383 4751 scope.go:117] "RemoveContainer" containerID="d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.721888 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.748073 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.768119 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.780912 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.790050 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.800422 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.811359 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.822406 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.935424 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.935499 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:25 crc kubenswrapper[4751]: I1002 10:52:25.935542 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.935673 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.935739 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:26.935718957 +0000 UTC m=+28.989945427 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.935805 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:26.935796259 +0000 UTC m=+28.990022719 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.935844 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:25 crc kubenswrapper[4751]: E1002 10:52:25.935873 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:26.93586498 +0000 UTC m=+28.990091440 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.036526 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.036585 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036712 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036723 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036762 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036732 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036778 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036787 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036841 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:27.036823462 +0000 UTC m=+29.091049922 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.036872 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:27.036850652 +0000 UTC m=+29.091077112 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.726783 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa"} Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.726881 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b"} Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.726910 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8539cc47cb7cefa10113c2bec7ae704568f0c0e1b2342982ae71b861f05d2c3c"} Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.729494 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d"} Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.758901 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.781052 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.798714 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.820611 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.842829 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.860617 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.878767 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.893326 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.905147 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.906157 4751 scope.go:117] "RemoveContainer" containerID="d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6" Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.906390 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.907800 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.928446 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.947358 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.947580 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.947635 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:28.947596 +0000 UTC m=+31.001822510 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.947742 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.947740 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.947860 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.947901 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:28.947805086 +0000 UTC m=+31.002031576 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: E1002 10:52:26.948503 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:28.948479854 +0000 UTC m=+31.002706414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.951731 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.966993 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:26 crc kubenswrapper[4751]: I1002 10:52:26.981882 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.002221 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:26Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.048514 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.048573 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048697 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048697 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048717 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048729 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048735 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048742 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048790 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:29.048773997 +0000 UTC m=+31.103000457 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.048812 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:29.048803088 +0000 UTC m=+31.103029558 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.550028 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.550076 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.550246 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:27 crc kubenswrapper[4751]: I1002 10:52:27.550289 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.550416 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:27 crc kubenswrapper[4751]: E1002 10:52:27.550566 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:28 crc kubenswrapper[4751]: I1002 10:52:28.966689 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:28 crc kubenswrapper[4751]: I1002 10:52:28.966866 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:28 crc kubenswrapper[4751]: E1002 10:52:28.966936 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:32.966894798 +0000 UTC m=+35.021121298 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:28 crc kubenswrapper[4751]: E1002 10:52:28.967047 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:28 crc kubenswrapper[4751]: I1002 10:52:28.967057 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:28 crc kubenswrapper[4751]: E1002 10:52:28.967142 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:32.967110844 +0000 UTC m=+35.021337334 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:28 crc kubenswrapper[4751]: E1002 10:52:28.967222 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:28 crc kubenswrapper[4751]: E1002 10:52:28.967315 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:32.967294159 +0000 UTC m=+35.021520669 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.068263 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.068337 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068485 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068508 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068524 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068525 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068564 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068585 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068594 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:33.068573539 +0000 UTC m=+35.122800009 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.068666 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:33.06864427 +0000 UTC m=+35.122870760 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.549844 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.549861 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.550263 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.550304 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.550390 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:29 crc kubenswrapper[4751]: E1002 10:52:29.550489 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.569073 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.584003 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.599240 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.615322 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.629547 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.644474 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.658399 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.743853 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f"} Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.757222 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.772658 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.785290 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.797448 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.810082 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.822980 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:29 crc kubenswrapper[4751]: I1002 10:52:29.836483 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.815752 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gvqc8"] Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.816406 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zzb9w"] Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.816539 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.816609 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.818743 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.819076 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.819285 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.819554 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.819709 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.819935 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.820027 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.839479 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.855491 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.866942 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.880099 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.884030 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js2j5\" (UniqueName: \"kubernetes.io/projected/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-kube-api-access-js2j5\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.884099 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-host\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.884117 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-serviceca\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.884212 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb39af81-e656-4533-8b2e-9058f44dc246-hosts-file\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.884363 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvgn9\" (UniqueName: \"kubernetes.io/projected/bb39af81-e656-4533-8b2e-9058f44dc246-kube-api-access-wvgn9\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.891947 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.902287 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.912408 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.931755 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.954384 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.965880 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.978026 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.978103 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.982221 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985501 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-host\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985557 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-serviceca\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985576 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb39af81-e656-4533-8b2e-9058f44dc246-hosts-file\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985589 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-host\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985594 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvgn9\" (UniqueName: \"kubernetes.io/projected/bb39af81-e656-4533-8b2e-9058f44dc246-kube-api-access-wvgn9\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985665 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js2j5\" (UniqueName: \"kubernetes.io/projected/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-kube-api-access-js2j5\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.985696 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb39af81-e656-4533-8b2e-9058f44dc246-hosts-file\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.987924 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-serviceca\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:30 crc kubenswrapper[4751]: I1002 10:52:30.988451 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.007782 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js2j5\" (UniqueName: \"kubernetes.io/projected/5f3d053f-7d6b-4fbb-88b3-e643f455bae7-kube-api-access-js2j5\") pod \"node-ca-gvqc8\" (UID: \"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\") " pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.007924 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvgn9\" (UniqueName: \"kubernetes.io/projected/bb39af81-e656-4533-8b2e-9058f44dc246-kube-api-access-wvgn9\") pod \"node-resolver-zzb9w\" (UID: \"bb39af81-e656-4533-8b2e-9058f44dc246\") " pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.012988 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.045885 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.059225 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.079366 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.089676 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.098465 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.112247 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.122906 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.131646 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gvqc8" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.135009 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.138785 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zzb9w" Oct 02 10:52:31 crc kubenswrapper[4751]: W1002 10:52:31.140930 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f3d053f_7d6b_4fbb_88b3_e643f455bae7.slice/crio-ae824999a1b8a53705ea7f476719c46714669db9f0b64b7de3b6ad74089edced WatchSource:0}: Error finding container ae824999a1b8a53705ea7f476719c46714669db9f0b64b7de3b6ad74089edced: Status 404 returned error can't find the container with id ae824999a1b8a53705ea7f476719c46714669db9f0b64b7de3b6ad74089edced Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.145343 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.160385 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.171834 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.184774 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.193408 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-b4wsd"] Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.193819 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.195461 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.196107 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.196432 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.197392 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.197929 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.201555 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.221482 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.240673 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.253518 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.264221 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.276535 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.287584 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.287944 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b26gd\" (UniqueName: \"kubernetes.io/projected/173e3cf6-4beb-4df2-a3f8-80504ed406cc-kube-api-access-b26gd\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.287988 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/173e3cf6-4beb-4df2-a3f8-80504ed406cc-rootfs\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.288014 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/173e3cf6-4beb-4df2-a3f8-80504ed406cc-mcd-auth-proxy-config\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.288036 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/173e3cf6-4beb-4df2-a3f8-80504ed406cc-proxy-tls\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.297097 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.307788 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.325258 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.337013 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.345912 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.362452 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.383997 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.389431 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/173e3cf6-4beb-4df2-a3f8-80504ed406cc-proxy-tls\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.389462 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/173e3cf6-4beb-4df2-a3f8-80504ed406cc-mcd-auth-proxy-config\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.389499 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b26gd\" (UniqueName: \"kubernetes.io/projected/173e3cf6-4beb-4df2-a3f8-80504ed406cc-kube-api-access-b26gd\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.389515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/173e3cf6-4beb-4df2-a3f8-80504ed406cc-rootfs\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.389568 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/173e3cf6-4beb-4df2-a3f8-80504ed406cc-rootfs\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.390420 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/173e3cf6-4beb-4df2-a3f8-80504ed406cc-mcd-auth-proxy-config\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.392975 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/173e3cf6-4beb-4df2-a3f8-80504ed406cc-proxy-tls\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.393691 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.403332 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b26gd\" (UniqueName: \"kubernetes.io/projected/173e3cf6-4beb-4df2-a3f8-80504ed406cc-kube-api-access-b26gd\") pod \"machine-config-daemon-b4wsd\" (UID: \"173e3cf6-4beb-4df2-a3f8-80504ed406cc\") " pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.506109 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:52:31 crc kubenswrapper[4751]: W1002 10:52:31.517882 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod173e3cf6_4beb_4df2_a3f8_80504ed406cc.slice/crio-20f9fe08d85d4266cad877468c5267c51b665e0f4e2fc1ff84156d05ee6e6019 WatchSource:0}: Error finding container 20f9fe08d85d4266cad877468c5267c51b665e0f4e2fc1ff84156d05ee6e6019: Status 404 returned error can't find the container with id 20f9fe08d85d4266cad877468c5267c51b665e0f4e2fc1ff84156d05ee6e6019 Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.549520 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.549535 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.549557 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.550016 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.550242 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.550342 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.600081 4751 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.601619 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.601662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.601671 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.601730 4751 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.605539 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pccq9"] Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.606454 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.612820 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8vj4p"] Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.613272 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.618937 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.619009 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.619027 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.619302 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-7h88l"] Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.619971 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623451 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623547 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623677 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623686 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623733 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623760 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.623818 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.628441 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.628471 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.628451 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.629931 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.639653 4751 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.639877 4751 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.643616 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.644486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.644571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.644661 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.644743 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.661466 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.690565 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691097 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-os-release\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691367 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691488 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691613 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-conf-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691832 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.691959 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-bin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692085 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-binary-copy\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692221 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-multus-certs\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692350 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692462 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-etc-kubernetes\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692579 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692708 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692836 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgl42\" (UniqueName: \"kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.692957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-daemon-config\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693072 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693200 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693307 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693447 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693677 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h75lw\" (UniqueName: \"kubernetes.io/projected/5e66e701-8712-4399-b2a8-b42e8cf716a3-kube-api-access-h75lw\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693797 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-system-cni-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.693887 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694011 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-cni-binary-copy\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694139 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-hostroot\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694318 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694465 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-cnibin\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694577 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694676 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694801 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-netns\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.694928 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-multus\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695040 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-kubelet\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695125 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695252 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695432 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-system-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695515 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-cnibin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695576 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-k8s-cni-cncf-io\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695720 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wng8h\" (UniqueName: \"kubernetes.io/projected/bdca1606-9dd6-472f-bb38-4f2ac994f820-kube-api-access-wng8h\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.695979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.696109 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.696229 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.696338 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-os-release\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.696436 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-socket-dir-parent\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.703637 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.707045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.707071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.707079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.707091 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.707100 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.719119 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.728391 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.733577 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.733609 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.733618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.733632 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.733643 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.736136 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.749709 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.749750 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.749760 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"20f9fe08d85d4266cad877468c5267c51b665e0f4e2fc1ff84156d05ee6e6019"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.750986 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zzb9w" event={"ID":"bb39af81-e656-4533-8b2e-9058f44dc246","Type":"ContainerStarted","Data":"f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.751051 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zzb9w" event={"ID":"bb39af81-e656-4533-8b2e-9058f44dc246","Type":"ContainerStarted","Data":"39f2039edf19bf5b7cdd58832e17aad2642a7c9492bd9ad4658b736c98024c91"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.752156 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gvqc8" event={"ID":"5f3d053f-7d6b-4fbb-88b3-e643f455bae7","Type":"ContainerStarted","Data":"187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.752201 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gvqc8" event={"ID":"5f3d053f-7d6b-4fbb-88b3-e643f455bae7","Type":"ContainerStarted","Data":"ae824999a1b8a53705ea7f476719c46714669db9f0b64b7de3b6ad74089edced"} Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.753232 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756344 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756378 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.756935 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.769249 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.774115 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.774162 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.774190 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.774208 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.774218 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.789159 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.793089 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: E1002 10:52:31.793214 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.794856 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.794890 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.794900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.794914 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.794922 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797311 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-system-cni-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797343 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797362 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-cni-binary-copy\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-hostroot\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797391 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797414 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-cnibin\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797429 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797443 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797443 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797459 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-netns\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797475 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-multus\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797474 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-system-cni-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797492 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-kubelet\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797534 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-kubelet\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797535 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797589 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797613 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797625 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-netns\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797635 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-system-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797632 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-hostroot\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797661 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797593 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797680 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-cnibin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797649 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-cnibin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797707 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-multus\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797712 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797723 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-k8s-cni-cncf-io\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797740 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797736 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797773 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-k8s-cni-cncf-io\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797749 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wng8h\" (UniqueName: \"kubernetes.io/projected/bdca1606-9dd6-472f-bb38-4f2ac994f820-kube-api-access-wng8h\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797809 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-system-cni-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797816 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797840 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797859 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797874 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-os-release\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-socket-dir-parent\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797919 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-os-release\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797935 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797952 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797968 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-conf-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.797985 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798000 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798016 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-bin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798016 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798020 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798041 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798042 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-binary-copy\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798076 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-multus-certs\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798085 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-os-release\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798094 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798111 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-etc-kubernetes\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798128 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798160 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798193 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgl42\" (UniqueName: \"kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798210 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-daemon-config\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798226 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798242 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798257 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798265 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-cni-binary-copy\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798292 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798272 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798325 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798339 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-socket-dir-parent\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798348 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h75lw\" (UniqueName: \"kubernetes.io/projected/5e66e701-8712-4399-b2a8-b42e8cf716a3-kube-api-access-h75lw\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798381 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-os-release\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798403 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-run-multus-certs\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798427 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798448 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-etc-kubernetes\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798460 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798468 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798491 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798500 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798572 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798587 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798615 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798628 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-binary-copy\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798670 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798699 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-conf-dir\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798702 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e66e701-8712-4399-b2a8-b42e8cf716a3-host-var-lib-cni-bin\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.798914 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bdca1606-9dd6-472f-bb38-4f2ac994f820-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.799161 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.799159 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5e66e701-8712-4399-b2a8-b42e8cf716a3-multus-daemon-config\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.799207 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bdca1606-9dd6-472f-bb38-4f2ac994f820-cnibin\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.803662 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.809481 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.816146 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h75lw\" (UniqueName: \"kubernetes.io/projected/5e66e701-8712-4399-b2a8-b42e8cf716a3-kube-api-access-h75lw\") pod \"multus-8vj4p\" (UID: \"5e66e701-8712-4399-b2a8-b42e8cf716a3\") " pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.816223 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wng8h\" (UniqueName: \"kubernetes.io/projected/bdca1606-9dd6-472f-bb38-4f2ac994f820-kube-api-access-wng8h\") pod \"multus-additional-cni-plugins-7h88l\" (UID: \"bdca1606-9dd6-472f-bb38-4f2ac994f820\") " pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.819670 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgl42\" (UniqueName: \"kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42\") pod \"ovnkube-node-pccq9\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.822382 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.841715 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.853356 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.864739 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.874118 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.891235 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.896961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.897016 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.897024 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.897038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.897047 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.905477 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.917432 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.919679 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.928830 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8vj4p" Oct 02 10:52:31 crc kubenswrapper[4751]: W1002 10:52:31.929612 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37ba8a11_317d_406b_8628_fcd43b164dbd.slice/crio-14a180816373f064036945a715f84b6144ed14b8a988f0604480e57cee77451a WatchSource:0}: Error finding container 14a180816373f064036945a715f84b6144ed14b8a988f0604480e57cee77451a: Status 404 returned error can't find the container with id 14a180816373f064036945a715f84b6144ed14b8a988f0604480e57cee77451a Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.930834 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: W1002 10:52:31.940340 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e66e701_8712_4399_b2a8_b42e8cf716a3.slice/crio-c0c1ac888193b10c44a70778c31c333a56361e05f2dfa7365281ebb3e8dc7cf0 WatchSource:0}: Error finding container c0c1ac888193b10c44a70778c31c333a56361e05f2dfa7365281ebb3e8dc7cf0: Status 404 returned error can't find the container with id c0c1ac888193b10c44a70778c31c333a56361e05f2dfa7365281ebb3e8dc7cf0 Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.942468 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7h88l" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.944633 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.959249 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: W1002 10:52:31.969815 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdca1606_9dd6_472f_bb38_4f2ac994f820.slice/crio-d853becd9e77f6aed3970920f0ca0781b94d155ad69ff152a014f2a234814004 WatchSource:0}: Error finding container d853becd9e77f6aed3970920f0ca0781b94d155ad69ff152a014f2a234814004: Status 404 returned error can't find the container with id d853becd9e77f6aed3970920f0ca0781b94d155ad69ff152a014f2a234814004 Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.976239 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.993492 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:31Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.999919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:31 crc kubenswrapper[4751]: I1002 10:52:31.999960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:31.999971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:31.999991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.000004 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:31Z","lastTransitionTime":"2025-10-02T10:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.002586 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.021294 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.040277 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.052894 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.088867 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.102604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.102647 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.102659 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.102675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.102687 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.127038 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.204895 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.204955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.204974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.204998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.205013 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.307204 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.307253 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.307264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.307278 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.307288 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.409766 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.409817 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.409828 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.409843 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.409853 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.511685 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.511734 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.511746 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.511766 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.511778 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.614618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.614672 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.614689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.614712 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.614730 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.718244 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.718285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.718298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.718316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.718329 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.758869 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" exitCode=0 Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.758951 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.758985 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"14a180816373f064036945a715f84b6144ed14b8a988f0604480e57cee77451a"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.761221 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80" exitCode=0 Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.761299 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.761325 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerStarted","Data":"d853becd9e77f6aed3970920f0ca0781b94d155ad69ff152a014f2a234814004"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.762984 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerStarted","Data":"6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.763020 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerStarted","Data":"c0c1ac888193b10c44a70778c31c333a56361e05f2dfa7365281ebb3e8dc7cf0"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.776566 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.792785 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.810601 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.821478 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.821524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.821537 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.821554 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.821565 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.825122 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.841909 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.852722 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.869570 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.882540 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.895007 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.911028 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.924745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.924783 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.924793 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.924810 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.924822 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:32Z","lastTransitionTime":"2025-10-02T10:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.925425 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.939128 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.951850 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.964093 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.979530 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:32 crc kubenswrapper[4751]: I1002 10:52:32.992141 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:32Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.003397 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.011289 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.011458 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.011499 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.011564 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.011641 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:41.011579287 +0000 UTC m=+43.065805727 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.011647 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.011762 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:41.011665109 +0000 UTC m=+43.065891559 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.011775 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:41.011769312 +0000 UTC m=+43.065995762 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.016953 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.025476 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.028517 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.028552 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.028560 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.028574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.028583 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.043129 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.065304 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.087153 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.105035 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.112078 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.112117 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112249 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112286 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112253 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112296 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112309 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112319 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112358 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:41.112339773 +0000 UTC m=+43.166566223 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.112374 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:41.112368384 +0000 UTC m=+43.166594834 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.117986 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.130512 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.130562 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.130574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.130612 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.130625 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.133805 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.168186 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.207598 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.233090 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.233152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.233164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.233220 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.233233 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.245736 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.335559 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.335608 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.335622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.335657 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.335670 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.438726 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.439282 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.439309 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.439342 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.439364 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.546271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.546322 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.546335 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.546353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.546368 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.549585 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.549605 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.549599 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.549698 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.550046 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:33 crc kubenswrapper[4751]: E1002 10:52:33.550158 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.652770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.652833 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.652864 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.652897 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.652913 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.755142 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.755415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.755424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.755436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.755445 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.768535 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.768592 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.768608 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.768619 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.768632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.769996 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerStarted","Data":"cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.785184 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.797630 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.813889 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.827257 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.840481 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.852593 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.856982 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.857012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.857020 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.857033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.857042 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.866755 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.886247 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.902908 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.916802 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.931766 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.942913 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.959334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.959386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.959404 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.959425 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.959440 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:33Z","lastTransitionTime":"2025-10-02T10:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.960067 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:33 crc kubenswrapper[4751]: I1002 10:52:33.975472 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.062054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.062132 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.062163 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.062209 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.062227 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.164566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.164615 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.164628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.164646 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.164660 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.267286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.267325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.267333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.267348 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.267359 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.369953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.370012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.370028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.370048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.370062 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.472742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.472801 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.472815 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.472834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.472847 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.575355 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.575406 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.575419 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.575438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.575450 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.678522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.678584 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.678602 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.678626 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.678647 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.778625 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.780677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.780754 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.780781 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.780814 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.780837 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.781263 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6" exitCode=0 Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.781320 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.807614 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.830498 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.850961 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.870555 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.884416 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.884458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.884471 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.884487 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.884502 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.891276 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.908228 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.926044 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.943417 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.958203 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.981553 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.987741 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.987775 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.987785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.987801 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:34 crc kubenswrapper[4751]: I1002 10:52:34.987813 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:34Z","lastTransitionTime":"2025-10-02T10:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.001584 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.021981 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.032512 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.049112 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.090922 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.090967 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.090985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.091009 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.091027 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.194251 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.194308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.194321 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.194338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.194350 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.296762 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.296799 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.296810 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.296827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.296839 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.399996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.400038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.400050 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.400067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.400080 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.502710 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.502780 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.502798 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.502822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.502839 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.549425 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.549459 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:35 crc kubenswrapper[4751]: E1002 10:52:35.549629 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.549649 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:35 crc kubenswrapper[4751]: E1002 10:52:35.549778 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:35 crc kubenswrapper[4751]: E1002 10:52:35.549959 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.605232 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.605310 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.605336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.605365 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.605388 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.708130 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.708218 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.708232 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.708260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.708277 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.787037 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642" exitCode=0 Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.787108 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.810845 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.812710 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.812958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.813104 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.813319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.813483 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.833833 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.849147 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.869503 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.882276 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.893722 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.904705 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.914450 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.915312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.915349 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.915359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.915374 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.915387 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:35Z","lastTransitionTime":"2025-10-02T10:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.924594 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.936379 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.947721 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.960593 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.981559 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:35 crc kubenswrapper[4751]: I1002 10:52:35.996141 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:35Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.018527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.018561 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.018571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.018584 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.018593 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.122139 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.122208 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.122221 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.122238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.122250 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.224685 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.224720 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.224730 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.224742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.224751 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.327537 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.327597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.327616 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.327641 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.327659 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.429800 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.429853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.429865 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.429881 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.429895 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.533461 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.533531 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.533551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.533581 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.533600 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.636338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.636388 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.636402 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.636420 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.636434 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.739466 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.739502 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.739511 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.739526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.739536 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.795260 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.798897 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33" exitCode=0 Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.798958 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.816604 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.836560 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.842114 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.842152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.842185 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.842207 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.842229 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.860619 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.882709 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.897864 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.912849 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.926205 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.937004 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.945397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.945447 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.945465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.945489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.945506 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:36Z","lastTransitionTime":"2025-10-02T10:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.956659 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.970633 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:36 crc kubenswrapper[4751]: I1002 10:52:36.984747 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.000556 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:36Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.010130 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.023982 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.047887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.047935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.047951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.047971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.047986 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.150539 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.150607 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.150624 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.150646 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.150664 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.253471 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.253540 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.253561 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.253588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.253609 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.356208 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.356278 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.356298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.356327 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.356345 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.459782 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.459841 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.459857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.459882 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.459899 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.549917 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.549994 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.550039 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:37 crc kubenswrapper[4751]: E1002 10:52:37.550128 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:37 crc kubenswrapper[4751]: E1002 10:52:37.550380 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:37 crc kubenswrapper[4751]: E1002 10:52:37.550471 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.562892 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.562951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.563024 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.563049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.563065 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.666731 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.666769 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.666780 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.666796 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.666805 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.770456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.770516 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.770538 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.770567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.770589 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.807262 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerStarted","Data":"02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.831894 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.852387 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.872085 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.874565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.874618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.874649 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.874673 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.874690 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.891571 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.909974 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.940358 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.960556 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.977224 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.977286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.977306 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.977330 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.977346 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:37Z","lastTransitionTime":"2025-10-02T10:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.979404 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:37 crc kubenswrapper[4751]: I1002 10:52:37.998660 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:37Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.016903 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.037855 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.057567 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.075458 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.079508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.079541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.079551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.079567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.079578 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.090796 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.182383 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.182430 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.182443 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.182458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.182471 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.285907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.285958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.285974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.285997 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.286014 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.388456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.388497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.388510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.388526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.388538 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.490693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.490718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.490725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.490737 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.490745 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.592513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.592579 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.592599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.592629 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.592646 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.695623 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.695683 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.695700 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.695720 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.695735 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.798509 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.798628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.798656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.798689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.798713 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.815548 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657" exitCode=0 Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.815659 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.829996 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.830582 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.830691 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.841312 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.857845 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.868962 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.874480 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.885813 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901546 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901783 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901817 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901829 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901846 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.901858 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:38Z","lastTransitionTime":"2025-10-02T10:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.914468 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.928259 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.939523 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.957159 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.972431 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:38 crc kubenswrapper[4751]: I1002 10:52:38.985818 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:38Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.003094 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.004968 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.005018 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.005036 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.005058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.005078 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.015329 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.031303 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.043573 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.055388 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.070368 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.084336 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.102900 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.107711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.107767 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.107785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.107806 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.107823 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.117310 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.130596 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.142735 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.154008 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.171807 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.187318 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.200355 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.211065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.211123 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.211142 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.211198 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.211218 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.212568 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.239736 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.314605 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.314646 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.314654 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.314670 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.314678 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.417675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.417725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.417739 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.417760 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.417774 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.520979 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.521046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.521065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.521092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.521111 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.549889 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:39 crc kubenswrapper[4751]: E1002 10:52:39.550111 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.549942 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.550249 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:39 crc kubenswrapper[4751]: E1002 10:52:39.550447 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:39 crc kubenswrapper[4751]: E1002 10:52:39.550623 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.570051 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.591377 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.610471 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.623666 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.623715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.623727 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.623744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.623755 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.630412 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.651866 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.667269 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.682038 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.697688 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.708101 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.726255 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.726285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.726294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.726308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.726318 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.729896 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.747069 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.768532 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.786265 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.803272 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.829860 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.829919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.829937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.829959 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.829972 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.837158 4751 generic.go:334] "Generic (PLEG): container finished" podID="bdca1606-9dd6-472f-bb38-4f2ac994f820" containerID="ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91" exitCode=0 Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.837235 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerDied","Data":"ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.837817 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.863966 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.875870 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.885836 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.900542 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.925527 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.933807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.933857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.933873 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.933895 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.933911 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:39Z","lastTransitionTime":"2025-10-02T10:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.944983 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.958756 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.974641 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:39 crc kubenswrapper[4751]: I1002 10:52:39.992251 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.004526 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.017337 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.030051 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.043530 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.043726 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.043820 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.043920 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.044000 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.043974 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.064212 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.083366 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.095311 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.105679 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.116158 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.125124 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.133675 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.146664 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.146699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.146711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.146728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.146740 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.152886 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.175466 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.197822 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.214873 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.228474 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.245054 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.248265 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.248301 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.248311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.248326 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.248337 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.257870 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.268867 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.276354 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.350799 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.350993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.351080 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.351149 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.351231 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.455125 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.455329 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.455357 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.455388 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.455416 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.550490 4751 scope.go:117] "RemoveContainer" containerID="d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.558983 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.559043 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.559057 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.559074 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.559084 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.662707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.662755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.662771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.662793 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.662811 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.766033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.766066 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.766074 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.766089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.766098 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.844363 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.847654 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.848209 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.852810 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" event={"ID":"bdca1606-9dd6-472f-bb38-4f2ac994f820","Type":"ContainerStarted","Data":"10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.865545 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.868308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.868340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.868354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.868369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.868380 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.877340 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.895329 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.910666 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.925750 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.938771 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.950082 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.971321 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.971387 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.971411 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.971439 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.971462 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:40Z","lastTransitionTime":"2025-10-02T10:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.977556 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:40 crc kubenswrapper[4751]: I1002 10:52:40.991553 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:40Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.011904 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.027832 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.040515 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.061938 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.074684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.074742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.074753 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.074768 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.074779 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.081949 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.098529 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.102053 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.102256 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:52:57.102230059 +0000 UTC m=+59.156456529 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.102395 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.102534 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.102578 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.102738 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:57.102725853 +0000 UTC m=+59.156952303 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.102672 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.102917 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:57.102906327 +0000 UTC m=+59.157132777 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.111928 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.132093 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.153936 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.165240 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.175637 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.176879 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.176926 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.176941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.176958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.176971 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.186103 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.197251 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.203958 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.204013 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204107 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204122 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204131 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204183 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:57.204154756 +0000 UTC m=+59.258381206 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204448 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204459 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204468 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.204491 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:52:57.204483145 +0000 UTC m=+59.258709595 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.213795 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.226626 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.240522 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.252209 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.261872 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.275465 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.279674 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.279711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.279721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.279735 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.279746 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.382444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.382485 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.382498 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.382514 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.382522 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.485072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.485111 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.485120 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.485133 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.485142 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.549791 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.549898 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.549803 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.550030 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.549951 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:41 crc kubenswrapper[4751]: E1002 10:52:41.550231 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.586878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.586946 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.586964 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.586988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.587003 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.689774 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.689821 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.689832 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.689848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.689858 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.792909 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.792978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.792994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.793019 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.793036 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.858955 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/0.log" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.863074 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075" exitCode=1 Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.863224 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.864146 4751 scope.go:117] "RemoveContainer" containerID="3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.882908 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.895096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.895137 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.895151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.895192 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.895209 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.898194 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.916426 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.930162 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.945544 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.964131 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.974088 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.993967 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:41Z\\\",\\\"message\\\":\\\".Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:41.666451 6026 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:41.666470 6026 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 10:52:41.666475 6026 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 10:52:41.666495 6026 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 10:52:41.666506 6026 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 10:52:41.666510 6026 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:41.666755 6026 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:41.666812 6026 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:41.666827 6026 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 10:52:41.666833 6026 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 10:52:41.666857 6026 factory.go:656] Stopping watch factory\\\\nI1002 10:52:41.666870 6026 ovnkube.go:599] Stopped ovnkube\\\\nI1002 10:52:41.666888 6026 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:41.666891 6026 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 10:52:41.666908 6026 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:41Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.997383 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.997426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.997438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.997456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:41 crc kubenswrapper[4751]: I1002 10:52:41.997468 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:41Z","lastTransitionTime":"2025-10-02T10:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.014225 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.028210 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.043882 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.054940 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.075137 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.091722 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.099247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.099286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.099296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.099310 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.099324 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.179475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.179548 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.179575 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.179606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.179631 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.197954 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.202480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.202530 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.202545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.202568 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.202584 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.224653 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.228958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.228984 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.228994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.229007 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.229016 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.246608 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.250525 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.250603 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.250620 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.250638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.250675 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.272699 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.277020 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.277062 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.277076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.277094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.277107 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.301088 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: E1002 10:52:42.301267 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.302718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.302748 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.302760 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.302775 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.302786 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.405892 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.405949 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.405966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.405991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.406009 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.509035 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.509085 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.509099 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.509126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.509140 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.612047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.612090 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.612101 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.612117 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.612129 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.714747 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.714815 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.714832 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.714857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.714874 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.818025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.818096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.818139 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.818203 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.818236 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.870212 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/0.log" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.874460 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.875047 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.897859 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.918991 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.920660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.920708 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.920722 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.920744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.920762 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:42Z","lastTransitionTime":"2025-10-02T10:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.936486 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.961413 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.976142 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:42 crc kubenswrapper[4751]: I1002 10:52:42.989620 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:42Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.004432 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.023449 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.023515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.023538 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.023566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.023588 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.026507 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.047893 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.072252 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.093803 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.107706 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.126505 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.126564 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.126585 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.126610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.126629 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.139622 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:41Z\\\",\\\"message\\\":\\\".Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:41.666451 6026 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:41.666470 6026 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 10:52:41.666475 6026 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 10:52:41.666495 6026 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 10:52:41.666506 6026 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 10:52:41.666510 6026 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:41.666755 6026 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:41.666812 6026 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:41.666827 6026 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 10:52:41.666833 6026 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 10:52:41.666857 6026 factory.go:656] Stopping watch factory\\\\nI1002 10:52:41.666870 6026 ovnkube.go:599] Stopped ovnkube\\\\nI1002 10:52:41.666888 6026 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:41.666891 6026 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 10:52:41.666908 6026 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.161707 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.229507 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.229568 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.229589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.229614 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.229631 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.332047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.332109 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.332132 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.332159 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.332205 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.435119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.435589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.435736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.435900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.436059 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.539774 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.539875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.539894 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.539922 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.539940 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.550252 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:43 crc kubenswrapper[4751]: E1002 10:52:43.550392 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.550270 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.550423 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:43 crc kubenswrapper[4751]: E1002 10:52:43.550623 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:43 crc kubenswrapper[4751]: E1002 10:52:43.550757 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.643020 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.643085 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.643105 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.643133 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.643153 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.745876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.745982 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.746005 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.746033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.746054 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.849678 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.849746 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.849771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.849802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.849826 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.881747 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/1.log" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.883085 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/0.log" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.888056 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c" exitCode=1 Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.888140 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.888226 4751 scope.go:117] "RemoveContainer" containerID="3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.896027 4751 scope.go:117] "RemoveContainer" containerID="892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c" Oct 02 10:52:43 crc kubenswrapper[4751]: E1002 10:52:43.896879 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.921662 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.943164 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.954128 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.954312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.954331 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.954356 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.954389 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:43Z","lastTransitionTime":"2025-10-02T10:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.961456 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:43 crc kubenswrapper[4751]: I1002 10:52:43.988288 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:43Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.009891 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.027114 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.044329 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.058033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.058102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.058126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.058204 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.058231 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.061974 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.081724 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.083667 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz"] Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.084484 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.086873 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.087858 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.103334 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.124368 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.137489 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.138932 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.139016 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.139053 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.139086 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9zs\" (UniqueName: \"kubernetes.io/projected/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-kube-api-access-xc9zs\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.160998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.161048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.161067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.161090 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.161108 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.170810 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:41Z\\\",\\\"message\\\":\\\".Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:41.666451 6026 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:41.666470 6026 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 10:52:41.666475 6026 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 10:52:41.666495 6026 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 10:52:41.666506 6026 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 10:52:41.666510 6026 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:41.666755 6026 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:41.666812 6026 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:41.666827 6026 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 10:52:41.666833 6026 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 10:52:41.666857 6026 factory.go:656] Stopping watch factory\\\\nI1002 10:52:41.666870 6026 ovnkube.go:599] Stopped ovnkube\\\\nI1002 10:52:41.666888 6026 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:41.666891 6026 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 10:52:41.666908 6026 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.189680 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.208144 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.225620 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.240804 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.240929 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.240982 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.241026 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9zs\" (UniqueName: \"kubernetes.io/projected/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-kube-api-access-xc9zs\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.241714 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.241851 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.243468 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.248256 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.257627 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.264334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.264373 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.264408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.264427 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.264439 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.268628 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9zs\" (UniqueName: \"kubernetes.io/projected/4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d-kube-api-access-xc9zs\") pod \"ovnkube-control-plane-749d76644c-sg8vz\" (UID: \"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.273188 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.287617 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.312226 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:41Z\\\",\\\"message\\\":\\\".Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:41.666451 6026 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:41.666470 6026 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 10:52:41.666475 6026 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 10:52:41.666495 6026 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 10:52:41.666506 6026 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 10:52:41.666510 6026 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:41.666755 6026 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:41.666812 6026 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:41.666827 6026 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 10:52:41.666833 6026 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 10:52:41.666857 6026 factory.go:656] Stopping watch factory\\\\nI1002 10:52:41.666870 6026 ovnkube.go:599] Stopped ovnkube\\\\nI1002 10:52:41.666888 6026 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:41.666891 6026 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 10:52:41.666908 6026 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.331213 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.342860 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.355442 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.365659 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.366719 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.366751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.366763 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.366780 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.366794 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.380872 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.394058 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.405601 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.406148 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.418814 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: W1002 10:52:44.420260 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d9cba28_f92b_4fae_9df3_d67f7f4b1b4d.slice/crio-387621ff68fd547a38cc2b6170ba53f2ec994f3881ca1b1ff8307b80ed29ff3e WatchSource:0}: Error finding container 387621ff68fd547a38cc2b6170ba53f2ec994f3881ca1b1ff8307b80ed29ff3e: Status 404 returned error can't find the container with id 387621ff68fd547a38cc2b6170ba53f2ec994f3881ca1b1ff8307b80ed29ff3e Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.469484 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.469685 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.469763 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.469841 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.469901 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.573594 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.573637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.573648 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.573667 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.573680 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.676749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.676805 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.676816 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.676838 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.676851 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.780938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.781006 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.781024 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.781049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.781068 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.884425 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.884487 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.884508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.884535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.884552 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.896832 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" event={"ID":"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d","Type":"ContainerStarted","Data":"3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.896906 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" event={"ID":"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d","Type":"ContainerStarted","Data":"fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.896929 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" event={"ID":"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d","Type":"ContainerStarted","Data":"387621ff68fd547a38cc2b6170ba53f2ec994f3881ca1b1ff8307b80ed29ff3e"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.902239 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/1.log" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.910625 4751 scope.go:117] "RemoveContainer" containerID="892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c" Oct 02 10:52:44 crc kubenswrapper[4751]: E1002 10:52:44.910922 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.926565 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.950898 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.969215 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.986843 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.986921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.986945 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.986976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.986999 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:44Z","lastTransitionTime":"2025-10-02T10:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:44 crc kubenswrapper[4751]: I1002 10:52:44.994927 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.010769 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.025992 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.042412 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.055576 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.066195 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.089745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.089785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.089796 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.089812 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.089824 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.093461 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a5d4526c18ecbfd45bbd69f5701168de7b6846359a5947797d2e5a94bc74075\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:41Z\\\",\\\"message\\\":\\\".Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:41.666451 6026 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:41.666470 6026 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1002 10:52:41.666475 6026 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1002 10:52:41.666495 6026 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 10:52:41.666506 6026 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1002 10:52:41.666510 6026 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:41.666755 6026 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:41.666812 6026 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:41.666827 6026 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 10:52:41.666833 6026 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 10:52:41.666857 6026 factory.go:656] Stopping watch factory\\\\nI1002 10:52:41.666870 6026 ovnkube.go:599] Stopped ovnkube\\\\nI1002 10:52:41.666888 6026 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:41.666891 6026 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 10:52:41.666908 6026 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.109161 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.126477 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.143589 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.163694 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.178076 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.192943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.193013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.193030 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.193056 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.193071 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.200052 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.219390 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.234979 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.258071 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.275432 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.290977 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.295606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.295651 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.295660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.295675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.295685 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.308012 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.323722 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.341102 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.362481 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.380129 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.398882 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.398947 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.398966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.398992 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.399009 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.399294 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.414608 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.446510 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.465149 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.502652 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.502721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.502741 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.502769 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.502791 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.549080 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.549114 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.549241 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.549344 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.549475 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.549614 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.606520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.606622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.606651 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.606687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.606711 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.615312 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-tzdtd"] Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.615995 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.616095 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.638929 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.656449 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.656516 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8z4h\" (UniqueName: \"kubernetes.io/projected/278152b8-cf19-4211-95c9-2a162ff9549f-kube-api-access-v8z4h\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.660075 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.679397 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.699793 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.709299 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.709348 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.709360 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.709378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.709391 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.718958 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.738402 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.756382 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.757959 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.758044 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8z4h\" (UniqueName: \"kubernetes.io/projected/278152b8-cf19-4211-95c9-2a162ff9549f-kube-api-access-v8z4h\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.758246 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:45 crc kubenswrapper[4751]: E1002 10:52:45.758330 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:52:46.258303046 +0000 UTC m=+48.312529536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.778577 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.779792 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8z4h\" (UniqueName: \"kubernetes.io/projected/278152b8-cf19-4211-95c9-2a162ff9549f-kube-api-access-v8z4h\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.797435 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.812414 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.812479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.812502 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.812531 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.812559 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.822838 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.839751 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.856994 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.873865 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.895504 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.914994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.915038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.915049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.915068 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.915088 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:45Z","lastTransitionTime":"2025-10-02T10:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.919057 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:45 crc kubenswrapper[4751]: I1002 10:52:45.932076 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:45Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.017659 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.017761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.017787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.017819 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.017843 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.121065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.121136 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.121160 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.121229 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.121257 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.224955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.225019 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.225043 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.225086 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.225110 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.263722 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:46 crc kubenswrapper[4751]: E1002 10:52:46.263944 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:46 crc kubenswrapper[4751]: E1002 10:52:46.264092 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:52:47.264061708 +0000 UTC m=+49.318288188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.328432 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.328472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.328485 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.328502 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.328514 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.431930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.432005 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.432030 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.432062 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.432085 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.535040 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.535098 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.535116 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.535142 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.535164 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.638813 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.638875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.638892 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.638915 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.638932 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.742115 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.742231 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.742252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.742276 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.742294 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.845424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.845477 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.845494 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.845533 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.845551 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.948534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.948648 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.948671 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.948848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:46 crc kubenswrapper[4751]: I1002 10:52:46.948875 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:46Z","lastTransitionTime":"2025-10-02T10:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.051781 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.051866 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.051891 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.051921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.051943 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.155484 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.155565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.155588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.155618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.155639 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.259057 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.259121 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.259138 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.259161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.259218 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.273107 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.273383 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.273480 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:52:49.273453346 +0000 UTC m=+51.327679836 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.361868 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.361931 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.361948 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.361972 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.361989 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.466637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.466703 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.466728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.466759 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.466974 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.549211 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.549272 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.549338 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.549211 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.549414 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.549453 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.549581 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:47 crc kubenswrapper[4751]: E1002 10:52:47.549758 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.569460 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.569492 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.569500 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.569512 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.569521 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.672632 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.672697 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.672718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.672742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.672759 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.775631 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.775937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.775970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.776001 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.776025 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.879389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.879475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.879493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.879521 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.879539 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.982386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.982460 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.982482 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.982511 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:47 crc kubenswrapper[4751]: I1002 10:52:47.982536 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:47Z","lastTransitionTime":"2025-10-02T10:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.085988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.086043 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.086052 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.086066 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.086075 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.188513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.188569 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.188585 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.188606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.188622 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.291430 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.291505 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.291527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.291559 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.291588 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.394152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.394250 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.394285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.394319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.394341 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.497807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.497855 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.497867 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.497883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.497895 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.601656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.601718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.601728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.601750 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.601759 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.705715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.705789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.705819 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.705850 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.705869 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.809853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.809913 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.809923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.809943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.809956 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.913251 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.913587 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.913696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.913803 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:48 crc kubenswrapper[4751]: I1002 10:52:48.913886 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:48Z","lastTransitionTime":"2025-10-02T10:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.016483 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.016525 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.016539 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.016553 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.016561 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.118880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.118945 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.118961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.118988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.119005 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.221875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.221935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.221951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.221976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.221994 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.295093 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.295284 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.295371 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:52:53.295346494 +0000 UTC m=+55.349572954 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.324680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.324759 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.324771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.324789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.324855 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.427894 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.427994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.428012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.428036 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.428053 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.531125 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.531229 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.531252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.531277 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.531295 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.549676 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.549737 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.549690 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.549846 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.549933 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.550133 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.550272 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:49 crc kubenswrapper[4751]: E1002 10:52:49.550398 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.572954 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.588909 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.606699 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.622814 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.634720 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.634807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.634834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.634869 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.634898 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.644671 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.666790 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.685568 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.700246 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.727366 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.737475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.737532 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.737550 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.737574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.737592 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.747816 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.765478 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.782280 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.794893 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.811272 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.829063 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.840924 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.841013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.841025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.841046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.841058 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.846763 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.944273 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.944313 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.944325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.944343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:49 crc kubenswrapper[4751]: I1002 10:52:49.944355 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:49Z","lastTransitionTime":"2025-10-02T10:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.048037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.048523 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.048537 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.048562 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.048576 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.150907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.150953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.150964 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.150982 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.150995 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.254520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.254574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.254582 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.254599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.254611 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.357681 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.357736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.357749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.357767 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.357780 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.460875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.460937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.460955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.460980 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.460997 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.563571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.563628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.563664 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.563698 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.563722 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.665456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.665526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.665545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.665569 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.665588 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.768224 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.768265 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.768273 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.768287 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.768295 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.870684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.870725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.870737 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.870750 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.870758 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.973694 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.973770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.973784 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.973802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:50 crc kubenswrapper[4751]: I1002 10:52:50.973839 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:50Z","lastTransitionTime":"2025-10-02T10:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.076415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.076501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.076517 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.076543 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.076557 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.179161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.179228 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.179242 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.179260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.179273 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.282032 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.282079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.282089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.282108 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.282120 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.384797 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.384858 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.384876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.384900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.384921 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.487564 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.487617 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.487628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.487645 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.487656 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.549487 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.549543 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.549559 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.549487 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:51 crc kubenswrapper[4751]: E1002 10:52:51.549643 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:51 crc kubenswrapper[4751]: E1002 10:52:51.549764 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:51 crc kubenswrapper[4751]: E1002 10:52:51.549835 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:51 crc kubenswrapper[4751]: E1002 10:52:51.549882 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.589985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.590018 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.590028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.590043 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.590054 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.692694 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.692748 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.692766 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.692786 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.692802 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.796025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.796111 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.796135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.796163 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.796237 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.899536 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.899599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.899620 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.899643 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.899660 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:51Z","lastTransitionTime":"2025-10-02T10:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.941528 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.968370 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:51Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:51 crc kubenswrapper[4751]: I1002 10:52:51.985296 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:51Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:51.999929 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:51Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.002065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.002133 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.002150 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.002205 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.002226 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.016767 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.028408 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.054120 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.073047 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.092812 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.106448 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.106504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.106520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.106545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.106562 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.113435 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.130451 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.143401 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.153050 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.169132 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.190206 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.203958 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.208825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.208872 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.208889 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.208912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.208930 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.219812 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.312227 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.312264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.312274 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.312288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.312299 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.415725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.415777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.415793 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.415818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.415837 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.518323 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.518371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.518387 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.518409 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.518428 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.620769 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.620828 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.620840 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.620856 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.620869 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.677979 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.678028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.678046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.678067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.678078 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.697588 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.702358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.702399 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.702408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.702422 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.702432 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.720353 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.723862 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.723902 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.723917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.723939 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.723955 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.743735 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.747472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.747506 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.747515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.747557 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.747570 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.764435 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.768699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.768774 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.768794 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.768818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.768840 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.793065 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:52Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:52 crc kubenswrapper[4751]: E1002 10:52:52.793707 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.797598 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.797649 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.797661 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.797678 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.797692 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.900960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.901015 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.901031 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.901058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:52 crc kubenswrapper[4751]: I1002 10:52:52.901092 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:52Z","lastTransitionTime":"2025-10-02T10:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.004789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.004877 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.004905 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.004937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.004961 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.108394 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.108450 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.108522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.108573 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.108593 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.211584 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.211671 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.211689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.211712 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.211732 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.315128 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.315222 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.315246 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.315276 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.315300 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.342044 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.342358 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.342446 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:53:01.342423736 +0000 UTC m=+63.396650216 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.417919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.417977 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.417993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.418017 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.418034 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.520366 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.520440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.520463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.520497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.520522 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.549431 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.549460 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.549631 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.549663 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.549695 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.549861 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.549966 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:53 crc kubenswrapper[4751]: E1002 10:52:53.550113 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.623252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.623296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.623308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.623326 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.623338 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.725938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.725996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.726012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.726037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.726054 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.829288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.829376 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.829395 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.829424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.829443 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.932777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.932847 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.932871 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.932901 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:53 crc kubenswrapper[4751]: I1002 10:52:53.932921 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:53Z","lastTransitionTime":"2025-10-02T10:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.034901 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.034955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.034970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.034989 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.035004 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.078147 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.088687 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.101526 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.123288 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.137273 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.138463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.138518 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.138534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.138558 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.138606 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.160749 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.177332 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.194854 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.216311 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.235437 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.241301 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.241370 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.241392 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.241420 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.241443 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.250886 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.269910 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.285920 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.307090 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.323265 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.335316 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.348761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.348815 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.348827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.348846 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.348862 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.396434 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.409569 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:54Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.450825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.450866 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.450876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.450893 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.450904 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.554497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.554561 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.554578 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.554600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.554620 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.658272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.658345 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.658381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.658413 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.658435 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.761531 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.761637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.761658 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.761688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.761711 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.864965 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.865029 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.865041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.865063 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.865077 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.969111 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.969550 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.969799 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.969973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:54 crc kubenswrapper[4751]: I1002 10:52:54.970106 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:54Z","lastTransitionTime":"2025-10-02T10:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.073238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.073571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.073694 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.073807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.073914 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.176616 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.176662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.176676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.176691 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.176704 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.279485 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.279532 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.279547 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.279570 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.279587 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.382658 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.382714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.382736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.382762 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.382786 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.485809 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.485890 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.485912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.485941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.486111 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.549837 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.549899 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.549972 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:55 crc kubenswrapper[4751]: E1002 10:52:55.550070 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.550141 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:55 crc kubenswrapper[4751]: E1002 10:52:55.550309 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:55 crc kubenswrapper[4751]: E1002 10:52:55.550484 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:55 crc kubenswrapper[4751]: E1002 10:52:55.550564 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.589712 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.589790 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.589827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.589859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.589884 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.692885 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.692985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.693004 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.693028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.693046 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.796364 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.796428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.796447 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.796471 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.796488 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.899254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.899328 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.899346 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.899372 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:55 crc kubenswrapper[4751]: I1002 10:52:55.899392 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:55Z","lastTransitionTime":"2025-10-02T10:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.002089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.002156 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.002233 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.002264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.002284 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.105261 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.105326 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.105343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.105369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.105387 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.207827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.207878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.207889 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.207909 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.207926 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.310575 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.310645 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.310668 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.310699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.310721 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.413195 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.413239 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.413252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.413271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.413289 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.516424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.516493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.516518 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.516550 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.516574 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.619996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.620076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.620094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.620116 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.620134 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.723548 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.723639 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.723675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.723709 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.723733 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.827660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.827719 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.827737 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.827802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.827821 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.931299 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.931369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.931386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.931412 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:56 crc kubenswrapper[4751]: I1002 10:52:56.931429 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:56Z","lastTransitionTime":"2025-10-02T10:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.035221 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.035297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.035320 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.035349 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.035370 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.138499 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.138560 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.138577 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.138600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.138619 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.187975 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.188144 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:53:29.188115478 +0000 UTC m=+91.242341968 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.188726 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.188904 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.188996 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:53:29.188977941 +0000 UTC m=+91.243204421 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.189244 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.189350 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:53:29.189322041 +0000 UTC m=+91.243548531 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.189241 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.241296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.242032 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.242210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.242395 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.242535 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.291107 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.291264 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291416 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291449 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291468 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291494 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291530 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291548 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:53:29.291526225 +0000 UTC m=+91.345752705 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291552 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.291628 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:53:29.291602767 +0000 UTC m=+91.345829257 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.346098 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.346248 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.346278 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.346307 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.346327 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.448995 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.449037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.449047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.449065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.449079 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.550482 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.550697 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551310 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551432 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.551499 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551580 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.551596 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551888 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.551950 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: E1002 10:52:57.551969 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.654858 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.654918 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.654939 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.654966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.654987 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.757935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.758016 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.758042 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.758067 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.758086 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.861789 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.861863 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.861875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.861910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.861927 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.965240 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.965330 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.965350 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.965378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:57 crc kubenswrapper[4751]: I1002 10:52:57.965396 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:57Z","lastTransitionTime":"2025-10-02T10:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.068680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.068754 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.068766 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.068786 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.068801 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.171588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.171662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.171685 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.171716 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.171739 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.274417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.274472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.274489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.274510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.274527 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.377896 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.377981 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.378008 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.378045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.378068 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.481529 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.481587 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.481605 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.481629 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.481646 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.584302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.584375 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.584397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.584424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.584446 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.687249 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.687333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.687353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.687384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.687403 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.790667 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.790742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.790765 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.790795 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.790817 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.893352 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.893464 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.893482 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.893504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.893521 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.996289 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.996365 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.996389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.996418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:58 crc kubenswrapper[4751]: I1002 10:52:58.996441 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:58Z","lastTransitionTime":"2025-10-02T10:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.098593 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.098642 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.098678 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.098694 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.098732 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.201867 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.201948 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.201973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.202000 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.202018 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.305698 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.305750 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.305762 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.305780 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.305800 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.409243 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.409361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.409399 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.409441 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.409467 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.511817 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.511859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.511872 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.511889 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.511903 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.549359 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.549450 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.549375 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:52:59 crc kubenswrapper[4751]: E1002 10:52:59.549600 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.549629 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:52:59 crc kubenswrapper[4751]: E1002 10:52:59.549727 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:52:59 crc kubenswrapper[4751]: E1002 10:52:59.549862 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:52:59 crc kubenswrapper[4751]: E1002 10:52:59.550075 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.568749 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.593160 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.609088 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.614237 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.614292 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.614308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.614329 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.614344 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.632167 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.654653 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.669286 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.685073 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.700531 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.717459 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.717732 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.717906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.718034 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.718208 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.720918 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.735327 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.752049 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.767789 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.792494 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.808104 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821106 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.821327 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.836446 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.851444 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:52:59Z is after 2025-08-24T17:21:41Z" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.923454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.923492 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.923500 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.923515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:52:59 crc kubenswrapper[4751]: I1002 10:52:59.923527 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:52:59Z","lastTransitionTime":"2025-10-02T10:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.027344 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.027440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.027463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.027492 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.027515 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.130726 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.130822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.130836 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.130853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.130866 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.233631 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.233688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.233699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.233718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.233730 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.336082 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.336143 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.336153 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.336186 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.336196 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.439596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.439634 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.439643 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.439675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.439687 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.543512 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.543583 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.543602 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.543631 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.543650 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.550986 4751 scope.go:117] "RemoveContainer" containerID="892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.646813 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.647336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.647361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.647393 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.647416 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.750361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.750437 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.750453 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.750480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.750497 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.854325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.854384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.854400 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.854442 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.854460 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.957742 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.957807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.957827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.957853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.957877 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:00Z","lastTransitionTime":"2025-10-02T10:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.970546 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/1.log" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.977247 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc"} Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.978107 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:53:00 crc kubenswrapper[4751]: I1002 10:53:00.994329 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:00Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.015367 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.036707 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.055494 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.061010 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.061103 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.061123 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.061148 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.061165 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.074756 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.092517 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.114173 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.138510 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.158498 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.164302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.164369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.164390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.164431 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.164465 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.176332 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.209210 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.233279 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.259861 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.267513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.267589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.267613 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.267645 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.267670 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.284279 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.300587 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.324582 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.341452 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:01Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.370455 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.370493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.370504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.370520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.370531 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.437195 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.437329 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.437393 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:53:17.437380329 +0000 UTC m=+79.491606779 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.472840 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.473041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.473105 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.473188 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.473253 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.549584 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.549676 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.549705 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.550008 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.549881 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.550094 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.549778 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.550316 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.574904 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.574948 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.574960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.574975 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.574986 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.677770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.677802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.677811 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.677824 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.677832 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.780418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.780468 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.780481 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.780496 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.780507 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.883002 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.883044 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.883060 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.883080 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.883097 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.983357 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/2.log" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.984211 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/1.log" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.985636 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.985711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.985735 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.985764 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.985784 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:01Z","lastTransitionTime":"2025-10-02T10:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.989417 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" exitCode=1 Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.989466 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc"} Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.989527 4751 scope.go:117] "RemoveContainer" containerID="892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c" Oct 02 10:53:01 crc kubenswrapper[4751]: I1002 10:53:01.991781 4751 scope.go:117] "RemoveContainer" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" Oct 02 10:53:01 crc kubenswrapper[4751]: E1002 10:53:01.992248 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.017956 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.032798 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.047241 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.064577 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.088593 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.088676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.088700 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.088738 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.088763 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.095420 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://892e52b6d3a4ac958d27ebed7eb89cb9d085f8295c10d60bc2217835508bfa9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:52:42Z\\\",\\\"message\\\":\\\"Service (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 10:52:42.887711 6200 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 10:52:42.887734 6200 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 10:52:42.887767 6200 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 10:52:42.887788 6200 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 10:52:42.887791 6200 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 10:52:42.887809 6200 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 10:52:42.887817 6200 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 10:52:42.887871 6200 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 10:52:42.887903 6200 factory.go:656] Stopping watch factory\\\\nI1002 10:52:42.887917 6200 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 10:52:42.887932 6200 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 10:52:42.888018 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 10:52:42.888084 6200 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.119672 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.135490 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.155308 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.173831 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.187007 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.190966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.191022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.191036 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.191054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.191068 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.205077 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.217232 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.230091 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.245014 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.262850 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.276296 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.288592 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:02Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.293209 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.293254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.293267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.293286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.293298 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.396829 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.396875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.396885 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.396902 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.396913 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.499096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.499137 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.499145 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.499160 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.499188 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.602357 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.602417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.602434 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.602457 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.602477 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.705687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.705746 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.705771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.705798 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.705820 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.808912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.808978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.808993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.809014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.809029 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.911989 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.912048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.912064 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.912087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.912105 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:02Z","lastTransitionTime":"2025-10-02T10:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:02 crc kubenswrapper[4751]: I1002 10:53:02.996860 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/2.log" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.002367 4751 scope.go:117] "RemoveContainer" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.002746 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.014816 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.014867 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.014884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.014908 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.014925 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.021486 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.033520 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.044034 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.054476 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.068324 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.079420 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.096642 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.096707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.096727 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.096751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.096770 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.109794 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.112091 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.116215 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.116297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.116312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.116330 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.116343 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.124131 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.136301 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141235 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141281 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141314 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.141834 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.155067 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.162915 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.164391 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.164459 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.164484 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.164515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.164540 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.177500 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.182816 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.187255 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.187309 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.187329 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.187350 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.187365 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.194439 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.200675 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.200839 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.202522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.202580 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.202596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.202636 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.202649 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.210825 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.221992 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.236157 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.246807 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.255418 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:03Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.305339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.305382 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.305417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.305438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.305450 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.408825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.408872 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.408884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.408903 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.408917 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.511926 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.511975 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.511989 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.512009 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.512024 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.549543 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.549602 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.549670 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.549835 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.549964 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.549969 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.550127 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:03 crc kubenswrapper[4751]: E1002 10:53:03.550344 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.615216 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.615285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.615300 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.615319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.615332 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.717335 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.717375 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.717387 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.717401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.717412 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.819542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.819597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.819609 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.819624 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.819634 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.923343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.923718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.923745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.923777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:03 crc kubenswrapper[4751]: I1002 10:53:03.923798 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:03Z","lastTransitionTime":"2025-10-02T10:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.026635 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.026686 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.026698 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.026714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.026725 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.129207 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.129284 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.129311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.129339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.129360 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.231652 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.231697 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.231714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.231736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.231749 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.334362 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.334436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.334458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.334486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.334507 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.438049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.438094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.438109 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.438127 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.438139 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.540557 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.540603 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.540618 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.540634 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.540645 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.642489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.642554 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.642571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.642589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.642604 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.744479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.744547 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.744559 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.744574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.744585 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.847135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.847336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.847359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.847397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.847419 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.950281 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.950312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.950320 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.950331 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:04 crc kubenswrapper[4751]: I1002 10:53:04.950339 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:04Z","lastTransitionTime":"2025-10-02T10:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.052266 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.052296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.052304 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.052316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.052325 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.154094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.154129 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.154138 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.154152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.154161 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.257079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.257122 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.257133 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.257151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.257162 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.360030 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.360066 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.360075 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.360089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.360098 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.462207 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.462293 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.462317 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.462347 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.462371 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.549725 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.549813 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.549833 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.549943 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:05 crc kubenswrapper[4751]: E1002 10:53:05.550106 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:05 crc kubenswrapper[4751]: E1002 10:53:05.550284 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:05 crc kubenswrapper[4751]: E1002 10:53:05.550439 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:05 crc kubenswrapper[4751]: E1002 10:53:05.550552 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.564952 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.565157 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.565267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.565367 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.565436 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.668381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.668422 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.668434 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.668449 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.668460 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.771263 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.771304 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.771312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.771326 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.771337 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.873836 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.874051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.874113 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.874238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.874307 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.977040 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.977071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.977080 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.977092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:05 crc kubenswrapper[4751]: I1002 10:53:05.977100 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:05Z","lastTransitionTime":"2025-10-02T10:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.078914 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.078951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.078964 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.078979 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.078992 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.182237 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.182298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.182315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.182348 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.182366 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.285389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.285456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.285475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.285500 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.285522 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.388202 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.388227 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.388235 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.388247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.388256 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.490189 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.490232 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.490243 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.490260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.490271 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.592079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.592135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.592150 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.592194 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.592210 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.694410 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.694456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.694464 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.694479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.694488 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.797133 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.797251 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.797267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.797290 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.797309 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.899827 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.899859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.899868 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.899880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:06 crc kubenswrapper[4751]: I1002 10:53:06.899890 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:06Z","lastTransitionTime":"2025-10-02T10:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.003384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.003438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.003455 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.003479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.003497 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.106248 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.106295 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.106303 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.106318 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.106329 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.209353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.209409 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.209429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.209454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.209472 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.314689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.314775 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.314800 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.314830 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.314861 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.418349 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.418412 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.418437 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.418468 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.418490 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.520972 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.521012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.521026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.521041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.521052 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.549980 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.550050 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.550119 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:07 crc kubenswrapper[4751]: E1002 10:53:07.550267 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:07 crc kubenswrapper[4751]: E1002 10:53:07.551036 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.550863 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:07 crc kubenswrapper[4751]: E1002 10:53:07.551229 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:07 crc kubenswrapper[4751]: E1002 10:53:07.551420 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.623682 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.623719 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.623733 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.623749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.623760 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.726716 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.726767 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.726787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.726811 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.726830 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.829023 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.829061 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.829072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.829087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.829097 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.931588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.931634 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.931644 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.931662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:07 crc kubenswrapper[4751]: I1002 10:53:07.931671 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:07Z","lastTransitionTime":"2025-10-02T10:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.034088 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.034127 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.034138 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.034156 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.034186 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.137030 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.137080 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.137097 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.137121 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.137138 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.240379 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.240629 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.240641 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.240656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.240665 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.343935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.343994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.344019 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.344049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.344073 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.446751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.446816 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.446836 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.446862 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.446881 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.548680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.548779 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.548806 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.548834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.548853 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.650791 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.650842 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.650857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.650877 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.650893 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.753404 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.753454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.753471 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.753493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.753511 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.856908 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.856953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.856964 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.856980 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.856992 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.960230 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.960314 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.960333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.960359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:08 crc kubenswrapper[4751]: I1002 10:53:08.960376 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:08Z","lastTransitionTime":"2025-10-02T10:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.063162 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.063284 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.063310 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.063339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.063361 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.167123 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.167225 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.167260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.167290 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.167308 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.271151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.271207 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.271217 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.271231 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.271241 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.374151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.374283 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.374343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.374368 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.374385 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.477339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.477417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.477432 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.477456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.477472 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.549045 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.549124 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:09 crc kubenswrapper[4751]: E1002 10:53:09.549296 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.549312 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:09 crc kubenswrapper[4751]: E1002 10:53:09.549478 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.549557 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:09 crc kubenswrapper[4751]: E1002 10:53:09.549680 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:09 crc kubenswrapper[4751]: E1002 10:53:09.549847 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.562758 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.574523 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.579676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.579713 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.579722 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.579738 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.579751 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.593944 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.605513 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.617683 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.627786 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.635483 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.653764 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.666580 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.677165 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.681929 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.681978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.681990 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.682010 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.682021 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.688918 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.699537 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.708344 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.719650 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.728623 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.741488 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.751578 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:09Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.784328 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.784376 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.784392 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.784414 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.784436 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.886710 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.886744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.886755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.886770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.886780 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.989132 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.989270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.989294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.989321 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:09 crc kubenswrapper[4751]: I1002 10:53:09.989342 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:09Z","lastTransitionTime":"2025-10-02T10:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.091478 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.091578 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.091592 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.091609 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.091644 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.194146 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.194257 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.194279 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.194306 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.194328 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.297334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.297377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.297388 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.297404 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.297414 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.399215 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.399254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.399264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.399282 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.399296 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.502143 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.502217 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.502232 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.502254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.502270 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.604970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.605033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.605052 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.605076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.605094 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.706787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.706826 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.706836 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.706852 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.706864 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.812324 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.812358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.812367 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.812380 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.812389 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.914377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.914597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.914675 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.914781 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:10 crc kubenswrapper[4751]: I1002 10:53:10.914874 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:10Z","lastTransitionTime":"2025-10-02T10:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.017646 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.017680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.017693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.017708 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.017720 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.119630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.119682 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.119702 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.119725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.119742 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.222110 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.222151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.222180 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.222199 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.222217 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.324718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.324791 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.324803 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.324824 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.324837 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.426903 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.426958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.426973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.426993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.427008 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.529396 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.529489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.529513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.529535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.529605 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.549695 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.549723 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:11 crc kubenswrapper[4751]: E1002 10:53:11.549992 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.549784 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:11 crc kubenswrapper[4751]: E1002 10:53:11.550136 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.549726 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:11 crc kubenswrapper[4751]: E1002 10:53:11.550297 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:11 crc kubenswrapper[4751]: E1002 10:53:11.550383 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.631385 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.631441 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.631454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.631472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.631484 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.734788 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.735051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.735141 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.735267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.735349 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.838451 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.838958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.839058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.839147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.839244 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.942302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.942373 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.942390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.942415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:11 crc kubenswrapper[4751]: I1002 10:53:11.942432 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:11Z","lastTransitionTime":"2025-10-02T10:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.044711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.044746 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.044755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.044771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.044781 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.146404 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.146440 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.146452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.146469 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.146479 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.249200 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.249243 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.249256 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.249272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.249284 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.350953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.350984 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.350994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.351010 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.351020 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.454190 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.454233 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.454244 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.454263 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.454274 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.556247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.556304 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.556320 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.556343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.556358 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.659053 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.659131 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.659145 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.659162 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.659191 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.762333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.762403 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.762420 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.762444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.762468 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.866153 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.866216 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.866230 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.866247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.866260 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.968514 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.968563 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.968574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.968591 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:12 crc kubenswrapper[4751]: I1002 10:53:12.968620 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:12Z","lastTransitionTime":"2025-10-02T10:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.072303 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.072395 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.072419 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.072458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.072482 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.175159 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.175237 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.175250 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.175276 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.175291 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.278502 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.278574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.278594 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.278626 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.278648 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.381259 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.381323 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.381336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.381357 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.381373 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.484956 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.485029 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.485052 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.485083 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.485106 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.549909 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.550221 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.550249 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.550333 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.550701 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.550789 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.550961 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.551298 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.552705 4751 scope.go:117] "RemoveContainer" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.553142 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.560872 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.560944 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.560968 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.560998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.561022 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.580956 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:13Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.585681 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.585744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.585756 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.585783 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.585798 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.602019 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:13Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.608122 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.608253 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.608279 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.608312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.608338 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.625839 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:13Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.631151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.631258 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.631277 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.631302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.631321 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.649653 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:13Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.654402 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.654494 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.654542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.654569 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.654593 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.672203 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:13Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:13 crc kubenswrapper[4751]: E1002 10:53:13.672376 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.674659 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.674714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.674729 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.674755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.674771 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.777777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.777876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.777898 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.777933 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.777953 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.882087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.882408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.882479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.882545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.882617 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.985991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.986037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.986051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.986074 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:13 crc kubenswrapper[4751]: I1002 10:53:13.986086 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:13Z","lastTransitionTime":"2025-10-02T10:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.089398 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.089716 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.089821 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.089932 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.090018 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.192831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.192894 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.192921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.192950 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.192972 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.294711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.294741 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.294749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.294761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.294770 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.397571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.397639 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.397662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.397689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.397711 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.500980 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.501087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.501104 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.501127 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.501143 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.604680 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.604759 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.604787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.604820 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.604845 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.708425 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.708489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.708506 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.708540 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.708559 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.811541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.811611 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.811633 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.811653 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.811672 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.913767 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.913815 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.913831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.913854 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:14 crc kubenswrapper[4751]: I1002 10:53:14.913870 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:14Z","lastTransitionTime":"2025-10-02T10:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.016752 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.016876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.016948 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.016983 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.017052 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.119366 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.119425 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.119443 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.119463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.119477 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.222128 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.222210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.222228 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.222250 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.222267 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.325848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.325938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.325964 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.325993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.326040 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.429594 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.429660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.429676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.429699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.429715 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.532922 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.533006 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.533038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.533068 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.533092 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.549447 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.549566 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:15 crc kubenswrapper[4751]: E1002 10:53:15.549713 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.549797 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.549805 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:15 crc kubenswrapper[4751]: E1002 10:53:15.550022 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:15 crc kubenswrapper[4751]: E1002 10:53:15.550227 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:15 crc kubenswrapper[4751]: E1002 10:53:15.550348 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.635935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.636014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.636042 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.636079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.636105 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.738866 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.738922 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.738934 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.738954 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.738967 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.842622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.842666 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.842678 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.842696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.842709 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.945418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.945470 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.945488 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.945511 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:15 crc kubenswrapper[4751]: I1002 10:53:15.945530 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:15Z","lastTransitionTime":"2025-10-02T10:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.048869 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.048929 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.048947 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.048970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.048987 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.152784 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.152831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.152855 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.152878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.152916 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.255428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.255474 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.255486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.255501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.255515 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.358797 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.358883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.358907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.358933 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.358951 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.461804 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.461864 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.461879 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.461903 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.461919 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.564810 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.564847 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.564859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.564874 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.564886 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.668165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.668294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.668317 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.668347 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.668371 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.771293 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.771368 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.771392 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.771418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.771437 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.874386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.874430 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.874455 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.874473 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.874482 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.976788 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.976823 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.976831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.976848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:16 crc kubenswrapper[4751]: I1002 10:53:16.976857 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:16Z","lastTransitionTime":"2025-10-02T10:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.079841 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.079889 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.079906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.079921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.079932 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.182907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.182979 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.183005 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.183037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.183060 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.286264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.286335 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.286357 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.286383 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.286401 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.389110 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.389280 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.389315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.389340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.389357 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.492063 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.492118 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.492130 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.492146 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.492157 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.510079 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.510304 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.510421 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:53:49.510385376 +0000 UTC m=+111.564611876 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.549517 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.549669 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.549907 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.549972 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.550020 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.550267 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.550400 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:17 crc kubenswrapper[4751]: E1002 10:53:17.550586 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.594893 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.594975 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.594998 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.595023 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.595043 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.698284 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.698358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.698377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.698405 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.698425 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.801271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.801346 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.801371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.801436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.801457 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.903971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.904036 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.904056 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.904080 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:17 crc kubenswrapper[4751]: I1002 10:53:17.904097 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:17Z","lastTransitionTime":"2025-10-02T10:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.006951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.007003 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.007014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.007034 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.007047 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.110013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.110063 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.110072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.110092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.110106 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.212623 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.212686 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.212707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.212731 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.212749 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.315223 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.315281 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.315298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.315324 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.315343 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.419235 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.419319 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.419335 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.419360 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.419376 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.522100 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.522140 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.522151 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.522165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.522193 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.625361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.625419 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.625432 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.625452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.625469 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.727837 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.727891 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.727903 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.727917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.727926 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.830836 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.830900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.830962 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.830994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.831015 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.933679 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.933760 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.933785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.933817 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:18 crc kubenswrapper[4751]: I1002 10:53:18.933842 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:18Z","lastTransitionTime":"2025-10-02T10:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.036643 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.036682 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.036693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.036711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.036721 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.138539 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.138596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.138614 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.138637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.138653 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.241089 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.241142 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.241159 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.241207 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.241228 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.343839 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.343910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.344118 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.344146 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.344164 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.447877 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.447937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.447955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.447978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.447994 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.549083 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.549199 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:19 crc kubenswrapper[4751]: E1002 10:53:19.549248 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:19 crc kubenswrapper[4751]: E1002 10:53:19.549374 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.549432 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:19 crc kubenswrapper[4751]: E1002 10:53:19.549623 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.549749 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:19 crc kubenswrapper[4751]: E1002 10:53:19.549967 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.550896 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.550952 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.550970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.550993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.551011 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.574349 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.592457 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.608456 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.619595 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.629915 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.646033 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.653842 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.653873 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.653885 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.653904 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.653916 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.656895 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.667821 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.678036 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.685478 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.698424 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.709230 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.721096 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.733815 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.745406 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.756946 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.757225 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.757333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.757433 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.757536 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.759133 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.775923 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:19Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.860467 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.861148 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.861209 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.861241 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.861266 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.964239 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.964311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.964328 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.964353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:19 crc kubenswrapper[4751]: I1002 10:53:19.964412 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:19Z","lastTransitionTime":"2025-10-02T10:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.055387 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/0.log" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.055485 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e66e701-8712-4399-b2a8-b42e8cf716a3" containerID="6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8" exitCode=1 Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.055560 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerDied","Data":"6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.056394 4751 scope.go:117] "RemoveContainer" containerID="6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.067238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.067293 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.067311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.067334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.067352 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.074378 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.097703 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.114633 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.134644 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.151858 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.167327 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.169825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.169878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.169897 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.169923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.169941 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.190271 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.207307 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.224754 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.240155 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.271640 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.272788 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.272884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.272899 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.272917 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.272931 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.294326 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.309027 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.320191 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.331109 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.345645 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.355878 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:20Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.375480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.375526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.375543 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.375565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.375581 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.478658 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.478707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.478722 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.478743 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.478758 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.582271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.582339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.582361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.582386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.582405 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.685883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.685923 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.685934 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.685952 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.685964 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.789157 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.789253 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.789272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.789296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.789317 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.892250 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.892322 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.892337 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.892356 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.892367 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.995431 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.995508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.995524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.995551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:20 crc kubenswrapper[4751]: I1002 10:53:20.995569 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:20Z","lastTransitionTime":"2025-10-02T10:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.062466 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/0.log" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.062552 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerStarted","Data":"ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.078413 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.091234 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.097943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.097999 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.098016 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.098038 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.098055 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.109103 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.125505 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.138270 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.160387 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.181786 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.195677 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.201029 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.201065 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.201076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.201092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.201105 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.213379 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.229405 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.248154 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.263041 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.281973 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.302258 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.303610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.303686 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.303705 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.303728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.303741 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.317041 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.331543 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.344517 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:21Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.406396 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.406471 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.406497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.406524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.406545 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.510165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.510254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.510272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.510296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.510314 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.549118 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.549153 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.549162 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:21 crc kubenswrapper[4751]: E1002 10:53:21.549307 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.549371 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:21 crc kubenswrapper[4751]: E1002 10:53:21.549433 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:21 crc kubenswrapper[4751]: E1002 10:53:21.549603 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:21 crc kubenswrapper[4751]: E1002 10:53:21.549732 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.612609 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.612656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.612666 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.612692 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.612714 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.715881 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.715940 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.715956 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.715980 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.715997 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.819669 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.819728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.819745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.819771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.819789 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.922033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.922086 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.922103 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.922126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:21 crc kubenswrapper[4751]: I1002 10:53:21.922148 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:21Z","lastTransitionTime":"2025-10-02T10:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.024819 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.024865 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.024875 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.024893 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.024905 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.129943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.129983 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.129995 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.130012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.130023 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.232944 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.233002 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.233014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.233032 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.233044 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.335893 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.335969 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.335994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.336025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.336046 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.438383 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.438426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.438437 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.438454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.438466 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.541617 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.541912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.542105 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.542389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.542604 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.646049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.646114 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.646132 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.646157 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.646210 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.749498 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.749573 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.749598 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.749622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.749640 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.852455 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.852547 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.852565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.852591 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.852609 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.955522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.955574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.955589 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.955608 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:22 crc kubenswrapper[4751]: I1002 10:53:22.955635 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:22Z","lastTransitionTime":"2025-10-02T10:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.058043 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.058102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.058119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.058143 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.058164 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.161913 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.162044 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.162084 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.162113 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.162136 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.265023 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.265090 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.265110 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.265135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.265155 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.368040 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.368135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.368162 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.368226 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.368249 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.470305 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.470381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.470401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.470426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.470448 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.549413 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.549492 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.549659 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.549722 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.549738 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.549852 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.549929 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.550083 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.573163 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.573247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.573265 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.573290 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.573308 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.676535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.676607 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.676625 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.676649 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.676697 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.736498 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.736599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.736624 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.736656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.736684 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.758394 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:23Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.763831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.763971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.763995 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.764018 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.764068 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.786670 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:23Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.792122 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.792216 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.792234 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.792255 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.792303 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.812869 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:23Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.818359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.818419 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.818438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.818463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.818482 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.838878 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:23Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.844623 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.844682 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.844698 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.844719 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.844740 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.862977 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:23Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:23 crc kubenswrapper[4751]: E1002 10:53:23.863218 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.866761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.866848 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.866866 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.866891 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.866907 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.976606 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.976676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.976700 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.976732 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:23 crc kubenswrapper[4751]: I1002 10:53:23.976756 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:23Z","lastTransitionTime":"2025-10-02T10:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.079650 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.079718 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.079738 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.079764 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.079784 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.183392 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.183464 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.183486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.183518 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.183543 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.286847 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.286934 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.286961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.286993 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.287017 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.390113 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.390165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.390210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.390234 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.390254 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.493650 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.493734 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.493758 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.493785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.493803 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.597130 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.597309 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.597343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.597369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.597390 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.699785 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.699851 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.699877 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.699906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.699928 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.802765 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.802820 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.802830 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.802845 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.802857 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.905830 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.906118 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.906240 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.906338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:24 crc kubenswrapper[4751]: I1002 10:53:24.906413 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:24Z","lastTransitionTime":"2025-10-02T10:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.008225 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.008290 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.008317 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.008347 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.008365 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.111736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.111810 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.111834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.111865 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.111890 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.215070 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.215155 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.215524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.215841 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.215882 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.319260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.319359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.319426 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.319459 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.319529 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.422361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.422458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.422479 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.422505 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.422522 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.525696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.525786 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.525812 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.525843 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.525866 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.549572 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:25 crc kubenswrapper[4751]: E1002 10:53:25.549798 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.549594 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:25 crc kubenswrapper[4751]: E1002 10:53:25.549932 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.549598 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:25 crc kubenswrapper[4751]: E1002 10:53:25.550044 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.549571 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:25 crc kubenswrapper[4751]: E1002 10:53:25.550136 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.628354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.628417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.628439 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.628472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.628497 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.732054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.732119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.732137 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.732165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.732214 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.835402 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.835474 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.835497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.835527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.835549 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.938890 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.938953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.938971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.938996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:25 crc kubenswrapper[4751]: I1002 10:53:25.939015 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:25Z","lastTransitionTime":"2025-10-02T10:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.042041 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.042119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.042139 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.042209 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.042249 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.145318 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.145371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.145382 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.145401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.145412 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.248049 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.248120 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.248137 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.248208 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.248228 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.351522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.351599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.351637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.351671 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.351694 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.456723 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.457204 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.457386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.457545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.457742 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.561833 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.561900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.561918 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.561943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.561960 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.664503 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.664591 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.664615 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.664646 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.664669 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.767982 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.768039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.768052 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.768072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.768084 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.871503 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.871604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.871630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.871659 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.871681 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.974638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.974707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.974721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.974744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:26 crc kubenswrapper[4751]: I1002 10:53:26.974759 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:26Z","lastTransitionTime":"2025-10-02T10:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.077997 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.078081 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.078098 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.078123 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.078140 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.181620 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.181690 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.181728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.181760 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.181784 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.285491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.285603 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.285630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.285664 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.285689 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.389218 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.389285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.389303 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.389333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.389353 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.492872 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.492941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.492960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.493023 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.493040 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.550018 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.550044 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.550052 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.550115 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:27 crc kubenswrapper[4751]: E1002 10:53:27.550353 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:27 crc kubenswrapper[4751]: E1002 10:53:27.550513 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:27 crc kubenswrapper[4751]: E1002 10:53:27.550647 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:27 crc kubenswrapper[4751]: E1002 10:53:27.550734 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.551655 4751 scope.go:117] "RemoveContainer" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.596153 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.596270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.596286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.596305 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.596318 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.699143 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.699266 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.699291 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.699316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.699335 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.801973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.802027 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.802048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.802072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.802090 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.904447 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.904526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.904544 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.904571 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:27 crc kubenswrapper[4751]: I1002 10:53:27.904590 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:27Z","lastTransitionTime":"2025-10-02T10:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.008329 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.008393 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.008415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.008444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.008467 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.089594 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/2.log" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.092714 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.094508 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.110729 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.110782 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.110793 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.110810 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.110823 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.114283 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.128605 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.140456 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.155445 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.167471 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.189756 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.211292 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.213456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.213508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.213525 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.213546 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.213561 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.227226 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.247965 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.267523 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.284134 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.296828 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.312992 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.316325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.316374 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.316390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.316411 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.316427 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.328887 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.341340 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.351763 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.364994 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:28Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.418986 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.419026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.419037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.419053 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.419067 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.522130 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.522242 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.522269 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.522302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.522326 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.563884 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.625146 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.625312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.625334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.625359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.625376 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.728365 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.728415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.728433 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.728456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.728473 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.831749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.831801 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.831819 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.831845 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.831861 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.935389 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.935483 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.935524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.935560 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:28 crc kubenswrapper[4751]: I1002 10:53:28.935584 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:28Z","lastTransitionTime":"2025-10-02T10:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.038933 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.039002 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.039026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.039058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.039079 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.099693 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/3.log" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.101019 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/2.log" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.104772 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" exitCode=1 Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.104919 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.105018 4751 scope.go:117] "RemoveContainer" containerID="28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.107041 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.107422 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.126419 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.142068 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.142117 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.142136 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.142163 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.142224 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.150437 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.170151 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.190492 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.210447 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.228684 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.228936 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:33.228909727 +0000 UTC m=+155.283136217 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.229018 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.229066 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.229218 4751 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.229274 4751 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.229317 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:54:33.229295738 +0000 UTC m=+155.283522228 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.229363 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 10:54:33.229329939 +0000 UTC m=+155.283556419 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.231837 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.244752 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.244835 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.244859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.244884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.244904 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.253392 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.270450 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.295898 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.314845 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.329963 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.330104 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330214 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330256 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330279 4751 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330367 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 10:54:33.33033942 +0000 UTC m=+155.384565910 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330381 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330424 4751 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330464 4751 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.330547 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 10:54:33.330521185 +0000 UTC m=+155.384747665 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.345034 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:28Z\\\",\\\"message\\\":\\\"d as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.480341 6812 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.479155 6812 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 10:53:28.480750 6812 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 10:53:28.480839 6812 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.347943 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.348083 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.348166 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.348332 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.348367 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.387823 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.409664 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aefc404-7ddd-42cd-9ad0-1ee4004b0db2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.427233 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.444094 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.450712 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.450773 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.450782 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.450800 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.450813 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.458390 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.478884 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.490924 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.549796 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.549896 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.550062 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.550072 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.550114 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.550228 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.550444 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:29 crc kubenswrapper[4751]: E1002 10:53:29.550536 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.554226 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.554271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.554285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.554304 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.554316 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.563261 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aefc404-7ddd-42cd-9ad0-1ee4004b0db2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.586457 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.601002 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.616334 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.636163 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.653702 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.656608 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.656689 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.656713 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.656745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.656767 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.672548 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.695487 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.715465 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.735439 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.751090 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.759086 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.759125 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.759134 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.759147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.759156 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.764379 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.779509 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.792149 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.804227 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.813304 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.838855 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28ee954043f1485ac3b40eb39b1873d74b68c87ec8dffb4c841a8fbab97e64fc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:01Z\\\",\\\"message\\\":\\\"c-sg8vz in node crc\\\\nI1002 10:53:01.730337 6449 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-pccq9\\\\nI1002 10:53:01.730343 6449 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz after 0 failed attempt(s)\\\\nI1002 10:53:01.730318 6449 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:01.730352 6449 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz\\\\nI1002 10:53:01.730360 6449 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-pccq9 in node crc\\\\nI1002 10:53:01.730308 6449 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:28Z\\\",\\\"message\\\":\\\"d as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.480341 6812 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.479155 6812 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 10:53:28.480750 6812 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 10:53:28.480839 6812 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.853964 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:29Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.862092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.862153 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.862205 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.862238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.862261 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.965116 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.965218 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.965243 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.965270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:29 crc kubenswrapper[4751]: I1002 10:53:29.965291 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:29Z","lastTransitionTime":"2025-10-02T10:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.068822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.068870 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.068887 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.068910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.068930 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.114642 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/3.log" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.118354 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:53:30 crc kubenswrapper[4751]: E1002 10:53:30.118486 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.134977 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.150304 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aefc404-7ddd-42cd-9ad0-1ee4004b0db2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.171501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.171541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.171551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.171566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.171578 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.173004 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.199291 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.217146 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.241627 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.267379 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.273569 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.273602 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.273613 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.273628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.273640 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.285364 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.303012 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.320562 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.337379 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.369858 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:28Z\\\",\\\"message\\\":\\\"d as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.480341 6812 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.479155 6812 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 10:53:28.480750 6812 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 10:53:28.480839 6812 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.376344 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.376397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.376415 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.376439 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.376457 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.393296 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.414112 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.433761 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.454793 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.475050 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.479705 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.479777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.479795 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.479820 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.479838 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.493205 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:30Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.582941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.583005 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.583022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.583047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.583069 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.686247 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.686336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.686400 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.686433 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.686457 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.790166 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.790513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.790537 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.790567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.790587 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.893597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.893979 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.893997 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.894022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.894042 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.997701 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.997763 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.997783 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.997807 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:30 crc kubenswrapper[4751]: I1002 10:53:30.997825 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:30Z","lastTransitionTime":"2025-10-02T10:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.101377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.101486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.101506 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.101566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.101583 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.204135 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.204233 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.204255 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.204283 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.204303 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.307216 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.307286 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.307311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.307340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.307362 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.410960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.411033 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.411059 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.411088 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.411109 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.514283 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.514371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.514397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.514428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.514453 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.550065 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.550216 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.550103 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:31 crc kubenswrapper[4751]: E1002 10:53:31.550340 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.550425 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:31 crc kubenswrapper[4751]: E1002 10:53:31.550569 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:31 crc kubenswrapper[4751]: E1002 10:53:31.550786 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:31 crc kubenswrapper[4751]: E1002 10:53:31.551412 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.572250 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.618248 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.618316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.618338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.618364 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.618382 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.721880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.721953 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.721971 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.722009 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.722030 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.824729 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.824806 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.824830 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.824860 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.824885 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.927693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.927744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.927764 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.927825 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:31 crc kubenswrapper[4751]: I1002 10:53:31.927843 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:31Z","lastTransitionTime":"2025-10-02T10:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.030642 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.030706 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.030724 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.030754 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.030777 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.133270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.133311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.133324 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.133341 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.133353 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.236570 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.236626 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.236644 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.236667 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.236685 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.339478 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.339523 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.339535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.339552 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.339564 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.442758 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.442826 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.442847 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.442873 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.442892 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.546007 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.546077 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.546102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.546128 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.546235 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.649272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.649336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.649354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.649381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.649398 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.753156 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.753253 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.753270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.753298 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.753316 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.856240 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.856299 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.856316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.856341 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.856362 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.960583 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.960657 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.960677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.960702 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:32 crc kubenswrapper[4751]: I1002 10:53:32.960721 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:32Z","lastTransitionTime":"2025-10-02T10:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.065013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.065096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.065118 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.065144 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.065164 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.168602 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.168670 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.168693 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.168721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.168745 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.271262 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.271325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.271359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.271387 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.271408 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.374097 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.374161 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.374201 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.374229 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.374262 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.481096 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.481165 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.481196 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.481211 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.481227 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.549803 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.549898 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.549821 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.549953 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.550062 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.550102 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.550224 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.550361 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.584420 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.584491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.584514 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.584542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.584565 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.687223 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.687300 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.687314 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.687338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.687362 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.790092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.790147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.790164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.790223 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.790244 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.892933 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.893010 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.893037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.893070 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.893094 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.904214 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.904305 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.904361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.904385 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.904401 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.921385 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.926416 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.926454 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.926463 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.926478 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.926490 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.946831 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.952250 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.952311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.952334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.952367 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.952393 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.972039 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.976557 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.976645 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.976663 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.976715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:33 crc kubenswrapper[4751]: I1002 10:53:33.976734 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:33Z","lastTransitionTime":"2025-10-02T10:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:33 crc kubenswrapper[4751]: E1002 10:53:33.997604 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:33Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.002527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.002587 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.002610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.002638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.002661 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: E1002 10:53:34.021885 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:34Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:34 crc kubenswrapper[4751]: E1002 10:53:34.021993 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.023504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.023534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.023542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.023572 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.023582 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.126859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.126912 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.126930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.126954 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.126972 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.230597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.230657 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.230677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.230702 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.230720 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.334347 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.334401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.334421 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.334445 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.334463 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.437636 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.437712 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.437734 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.437765 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.437790 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.540664 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.540734 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.540752 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.540777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.540794 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.643857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.643928 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.643951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.643980 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.644001 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.747424 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.747487 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.747510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.747538 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.747560 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.858480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.858539 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.858557 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.858581 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.858599 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.961254 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.961578 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.961590 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.961610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:34 crc kubenswrapper[4751]: I1002 10:53:34.961630 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:34Z","lastTransitionTime":"2025-10-02T10:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.065012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.065069 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.065087 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.065108 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.065124 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.168976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.169077 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.169119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.169150 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.169201 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.272721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.272868 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.272965 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.273056 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.273107 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.376446 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.376510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.376527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.376552 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.376569 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.479659 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.479735 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.479753 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.479779 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.479796 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.549627 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.549713 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:35 crc kubenswrapper[4751]: E1002 10:53:35.549859 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.549967 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:35 crc kubenswrapper[4751]: E1002 10:53:35.550037 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.550046 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:35 crc kubenswrapper[4751]: E1002 10:53:35.550221 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:35 crc kubenswrapper[4751]: E1002 10:53:35.550356 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.582641 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.582699 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.582709 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.582725 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.582735 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.686160 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.686269 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.686288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.686312 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.686330 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.789672 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.789749 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.789768 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.789794 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.789812 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.892889 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.892996 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.893021 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.893047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.893065 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.997079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.997202 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.997221 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.997256 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:35 crc kubenswrapper[4751]: I1002 10:53:35.997275 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:35Z","lastTransitionTime":"2025-10-02T10:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.100301 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.100363 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.100378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.100400 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.100416 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.203381 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.203436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.203446 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.203460 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.203470 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.306871 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.306925 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.306986 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.307009 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.307025 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.410625 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.410684 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.410696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.410714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.410728 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.513858 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.513918 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.513941 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.513970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.513990 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.616615 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.616672 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.616690 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.616714 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.616733 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.720423 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.720498 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.720515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.720541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.720560 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.823895 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.823940 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.823955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.823974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.823987 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.926358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.926461 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.926480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.926504 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:36 crc kubenswrapper[4751]: I1002 10:53:36.926525 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:36Z","lastTransitionTime":"2025-10-02T10:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.029863 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.030325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.030482 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.030610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.030737 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.133963 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.134026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.134047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.134072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.134091 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.236755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.237126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.237371 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.237438 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.237464 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.342270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.342315 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.342325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.342340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.342349 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.444853 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.444959 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.444988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.445018 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.445047 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.549789 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.549831 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:37 crc kubenswrapper[4751]: E1002 10:53:37.549989 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.549789 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.549883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: E1002 10:53:37.550119 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.550154 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.550444 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.550456 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.550490 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.550509 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: E1002 10:53:37.550574 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:37 crc kubenswrapper[4751]: E1002 10:53:37.550355 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.659978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.660430 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.660840 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.660882 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.660907 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.763986 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.764051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.764069 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.764095 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.764113 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.867878 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.867976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.868007 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.868047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.868070 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.971017 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.971076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.971093 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.971117 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:37 crc kubenswrapper[4751]: I1002 10:53:37.971134 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:37Z","lastTransitionTime":"2025-10-02T10:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.074705 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.074770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.074787 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.074814 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.074831 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.177423 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.177477 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.177497 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.177522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.177542 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.281225 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.281283 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.281311 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.281334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.281350 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.384481 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.384531 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.384550 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.384572 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.384592 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.488540 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.488627 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.488651 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.488682 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.488704 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.592529 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.592600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.592613 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.592631 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.592641 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.696551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.696649 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.696674 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.696713 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.696736 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.800843 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.800924 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.800935 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.800958 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.800971 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.904936 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.905013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.905039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.905063 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:38 crc kubenswrapper[4751]: I1002 10:53:38.905082 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:38Z","lastTransitionTime":"2025-10-02T10:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.007752 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.007828 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.007852 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.007883 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.007907 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.112001 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.112149 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.112194 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.112261 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.112281 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.214892 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.214966 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.214985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.215012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.215031 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.318397 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.318444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.318466 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.318487 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.318502 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.421102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.421150 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.421199 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.421228 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.421250 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.524910 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.524970 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.524987 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.525012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.525031 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.549729 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.549794 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.549963 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:39 crc kubenswrapper[4751]: E1002 10:53:39.550099 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.550139 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:39 crc kubenswrapper[4751]: E1002 10:53:39.549942 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:39 crc kubenswrapper[4751]: E1002 10:53:39.550307 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:39 crc kubenswrapper[4751]: E1002 10:53:39.550442 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.570095 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.592514 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.610504 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.627408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.627491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.627526 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.627560 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.627585 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.639531 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.653813 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.664233 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aefc404-7ddd-42cd-9ad0-1ee4004b0db2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.673108 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.694127 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4558a8a3-d9af-4cb8-954a-ee57ebe76370\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2fce79d0c287c78276c423a4da66591c96c894ab527adcb72fc07bd2d1e0cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://053c003e68ab876f5ce044e94c140083a0387c5e2fb40f70eb57b96551b28cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bac1f5f3060b21b391cae56b30d714b25dd6e311c123523acaf77901d8e1354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede217fd4c602d51b6800d65fb9132b04b8ad8c0bc0e051016bae140e015321b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a29f9a94eda077e73d5b97cc2454a48ec497825258a0f78f5f4cddf7179fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.705588 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.718830 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.730637 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.730697 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.730715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.730757 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.730769 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.731848 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.753602 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.771791 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.791313 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.810977 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.826153 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.833268 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.833316 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.833332 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.833354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.833371 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.853158 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:28Z\\\",\\\"message\\\":\\\"d as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.480341 6812 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.479155 6812 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 10:53:28.480750 6812 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 10:53:28.480839 6812 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.871049 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.888331 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:39Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.935568 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.935622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.935640 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.935663 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:39 crc kubenswrapper[4751]: I1002 10:53:39.935680 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:39Z","lastTransitionTime":"2025-10-02T10:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.038674 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.038757 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.038777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.038802 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.038820 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.141921 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.141988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.142014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.142045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.142068 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.245801 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.245880 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.245904 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.245936 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.245959 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.348442 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.348529 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.348549 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.348574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.348591 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.451612 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.451696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.451733 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.451770 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.451788 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.555112 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.555204 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.555220 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.555260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.555272 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.659413 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.659496 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.659508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.659535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.659555 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.762955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.763037 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.763054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.763079 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.763099 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.866296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.866350 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.866366 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.866384 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.866397 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.969334 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.969422 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.969444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.969474 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:40 crc kubenswrapper[4751]: I1002 10:53:40.969496 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:40Z","lastTransitionTime":"2025-10-02T10:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.072968 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.073051 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.073074 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.073102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.073123 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.175565 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.175638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.175661 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.175692 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.175713 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.278755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.278824 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.278844 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.278873 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.278891 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.383101 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.383230 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.383275 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.383308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.383334 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.486763 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.486803 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.486844 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.486860 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.486869 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.549785 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.549836 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.549785 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.549912 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:41 crc kubenswrapper[4751]: E1002 10:53:41.549910 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:41 crc kubenswrapper[4751]: E1002 10:53:41.549995 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:41 crc kubenswrapper[4751]: E1002 10:53:41.550386 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:41 crc kubenswrapper[4751]: E1002 10:53:41.550517 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.589268 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.589336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.589355 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.589379 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.589396 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.691832 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.691896 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.691908 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.691930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.691944 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.795131 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.795233 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.795253 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.795292 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.795313 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.898536 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.898587 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.898600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.898619 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:41 crc kubenswrapper[4751]: I1002 10:53:41.898635 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:41Z","lastTransitionTime":"2025-10-02T10:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.001222 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.001296 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.001309 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.001324 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.001333 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.104524 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.104585 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.104604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.104630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.104651 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.207849 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.207926 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.207951 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.207983 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.208003 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.311489 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.311558 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.311584 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.311610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.311631 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.414156 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.414252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.414277 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.414302 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.414319 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.516988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.517055 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.517078 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.517106 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.517128 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.551519 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:53:42 crc kubenswrapper[4751]: E1002 10:53:42.551840 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.620422 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.620475 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.620496 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.620518 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.620538 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.723338 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.723417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.723436 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.723464 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.723484 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.827513 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.827583 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.827596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.827628 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.827641 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.932320 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.932377 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.932393 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.932417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:42 crc kubenswrapper[4751]: I1002 10:53:42.932433 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:42Z","lastTransitionTime":"2025-10-02T10:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.035462 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.035508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.035517 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.035534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.035545 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.138614 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.138686 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.138705 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.138735 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.138755 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.241541 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.241611 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.241629 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.241655 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.241673 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.345094 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.345157 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.345206 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.345231 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.345251 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.447991 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.448047 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.448064 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.448088 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.448105 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.549554 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.549658 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:43 crc kubenswrapper[4751]: E1002 10:53:43.549811 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.549931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:43 crc kubenswrapper[4751]: E1002 10:53:43.550270 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:43 crc kubenswrapper[4751]: E1002 10:53:43.550428 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.550482 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:43 crc kubenswrapper[4751]: E1002 10:53:43.550657 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.550930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.550973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.550985 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.551003 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.551013 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.654240 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.654460 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.654503 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.654535 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.654585 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.758891 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.759024 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.759050 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.759078 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.759138 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.862691 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.862782 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.862809 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.862838 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.862857 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.965884 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.965947 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.965967 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.965990 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:43 crc kubenswrapper[4751]: I1002 10:53:43.966009 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:43Z","lastTransitionTime":"2025-10-02T10:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.069893 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.070014 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.070039 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.070072 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.070093 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.124239 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.124331 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.124356 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.124392 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.124419 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.145866 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.151265 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.151341 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.151359 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.151386 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.151409 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.174984 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.179620 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.179688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.179709 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.179736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.179755 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.199367 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.205340 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.205417 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.205443 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.205483 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.205511 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.230056 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.236473 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.236551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.236570 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.236597 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.236615 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.258314 4751 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9439a9ad-297a-4aaf-997f-74421bdf3adb\\\",\\\"systemUUID\\\":\\\"8ae4f9d6-49b0-4b75-b40e-7ccb1b7c52da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:44Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:44 crc kubenswrapper[4751]: E1002 10:53:44.258551 4751 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.261262 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.261342 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.261369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.261403 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.261427 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.364125 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.364230 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.364260 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.364294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.364320 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.467444 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.467508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.467553 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.467580 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.467599 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.570839 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.570904 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.570919 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.570937 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.570949 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.673600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.673677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.673696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.673722 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.673740 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.782942 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.783008 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.783026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.783048 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.783066 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.885946 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.886028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.886045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.886071 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.886092 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.988928 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.989018 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.989044 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.989076 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:44 crc kubenswrapper[4751]: I1002 10:53:44.989099 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:44Z","lastTransitionTime":"2025-10-02T10:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.092771 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.092846 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.092871 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.092901 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.092922 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.195662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.195723 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.195741 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.195764 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.195782 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.299202 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.299242 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.299258 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.299281 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.299300 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.402272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.402337 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.402355 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.402380 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.402397 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.504894 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.504934 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.504955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.504974 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.504987 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.551244 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:45 crc kubenswrapper[4751]: E1002 10:53:45.551369 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.551244 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.551443 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:45 crc kubenswrapper[4751]: E1002 10:53:45.551500 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.551467 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:45 crc kubenswrapper[4751]: E1002 10:53:45.551647 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:45 crc kubenswrapper[4751]: E1002 10:53:45.551676 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.607297 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.607353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.607363 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.607378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.607392 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.710308 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.710818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.710886 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.710952 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.711011 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.813472 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.813534 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.813552 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.813603 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.813623 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.917599 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.917687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.917710 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.917739 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:45 crc kubenswrapper[4751]: I1002 10:53:45.917762 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:45Z","lastTransitionTime":"2025-10-02T10:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.020674 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.021288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.021542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.021748 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.021952 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.125139 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.125228 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.125249 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.125275 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.125291 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.227651 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.227711 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.227724 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.227743 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.227761 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.330465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.330510 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.330522 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.330542 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.330557 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.433341 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.433413 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.433427 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.433445 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.433458 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.536360 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.536401 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.536412 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.536428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.536442 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.638829 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.638906 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.638930 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.638961 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.638983 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.741819 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.741867 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.741886 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.741908 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.741925 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.845353 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.845416 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.845435 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.845457 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.845476 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.947592 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.947656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.947676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.947701 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:46 crc kubenswrapper[4751]: I1002 10:53:46.947724 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:46Z","lastTransitionTime":"2025-10-02T10:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.050726 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.050803 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.050815 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.050831 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.050843 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.154465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.154520 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.154532 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.154554 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.154567 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.257545 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.257619 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.257643 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.257676 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.257698 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.360613 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.360700 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.360720 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.360753 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.360773 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.463261 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.463314 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.463330 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.463351 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.463366 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.549918 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.550022 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:47 crc kubenswrapper[4751]: E1002 10:53:47.550101 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.550113 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.550134 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:47 crc kubenswrapper[4751]: E1002 10:53:47.550273 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:47 crc kubenswrapper[4751]: E1002 10:53:47.553727 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:47 crc kubenswrapper[4751]: E1002 10:53:47.554006 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.565657 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.565688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.565696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.565707 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.565716 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.668235 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.668294 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.668307 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.668325 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.668335 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.771652 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.771703 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.771721 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.771744 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.771763 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.874460 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.874527 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.874544 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.874570 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.874587 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.977343 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.977410 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.977428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.977452 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:47 crc kubenswrapper[4751]: I1002 10:53:47.977467 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:47Z","lastTransitionTime":"2025-10-02T10:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.080604 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.080690 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.080715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.080750 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.080775 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.184406 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.184465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.184480 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.184501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.184516 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.287406 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.287470 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.287490 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.287514 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.287532 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.390704 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.390761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.390777 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.390800 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.390817 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.494551 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.494621 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.494638 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.494662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.494681 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.598272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.598336 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.598361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.598390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.598409 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.701761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.701818 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.701834 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.701859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.701877 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.805164 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.805259 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.805285 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.805314 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.805337 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.908687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.908750 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.908768 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.908792 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:48 crc kubenswrapper[4751]: I1002 10:53:48.908809 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:48Z","lastTransitionTime":"2025-10-02T10:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.011768 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.011847 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.011870 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.011900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.011922 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.114701 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.114781 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.114798 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.114822 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.114839 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.217857 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.217938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.217962 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.217994 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.218018 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.321439 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.321515 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.321538 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.321567 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.321589 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.424224 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.424354 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.424385 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.424418 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.424453 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.527612 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.527683 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.527720 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.527753 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.527774 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.550032 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.550263 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.550528 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.550849 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.551077 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.551140 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.551333 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.551494 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.570993 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ecfe853e-665c-44f4-b5af-e068547fda04\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f151248697c43da396646f96f517c9115cdcf3c3ba45cbece6076c2f5c7b5ea2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d3b13db61cbf7e95c144739ccd96a6b3863d11b59c827120735f602406a746d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b530afb465c54bf078981e816be23c73f28791f139df14987fc57d0131251284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27f0636d553d48ee405836c4571c3c5b1d1380c61568d0dc703414066255b2d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.577663 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.577888 4751 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:49 crc kubenswrapper[4751]: E1002 10:53:49.577982 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs podName:278152b8-cf19-4211-95c9-2a162ff9549f nodeName:}" failed. No retries permitted until 2025-10-02 10:54:53.577954878 +0000 UTC m=+175.632181358 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs") pod "network-metrics-daemon-tzdtd" (UID: "278152b8-cf19-4211-95c9-2a162ff9549f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.592366 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.611394 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea59abab3850c6858b6144bcc274a1544b4b83c22ea7f17f985d65f7f6c9fd9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630477 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"173e3cf6-4beb-4df2-a3f8-80504ed406cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4529ea604ebe3a165a44a79c54030bd8b2cd13a5edc506cfbc38e80b60b20a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b26gd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b4wsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630593 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630632 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630644 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630662 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.630672 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.652260 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8vj4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e66e701-8712-4399-b2a8-b42e8cf716a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:53:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:19Z\\\",\\\"message\\\":\\\"2025-10-02T10:52:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21\\\\n2025-10-02T10:52:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_58fb7cff-e1d5-4beb-af5c-54edd7627a21 to /host/opt/cni/bin/\\\\n2025-10-02T10:52:34Z [verbose] multus-daemon started\\\\n2025-10-02T10:52:34Z [verbose] Readiness Indicator file check\\\\n2025-10-02T10:53:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:53:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h75lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8vj4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.671988 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b84dba9-7057-4569-88de-6074c7ad2ac3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5896e022187a85c34c9ba7033dc3938edefb9c7527f55b5defaef517031a58a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59d58828c84c1d7a4e42265256deeab157347ced7fbe2095795cbc2db40bb9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcef8540157058d84eff5f6cfefb8a889652dca2dfc0234a916124faf39746ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a680cb2de7596c3995e0332eba369012ae94ab42044b6424ac27cbb9c86ee55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.692016 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.713488 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3236de838e8afdca3dffe5b3ccf97622ddaa7ffc2d34bf093097035c20b2aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be7675f5ade3384b17383c28ee5fc50eb0489c95b79f870d32663aa680f6015b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.734204 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.734252 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.734264 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.734282 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.734296 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.735856 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.753149 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gvqc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3d053f-7d6b-4fbb-88b3-e643f455bae7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187c09cd98e34dcd04503596a5937a84cbdb4295dd5bea9bd9b06b45ecf5bb8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-js2j5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gvqc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.786323 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37ba8a11-317d-406b-8628-fcd43b164dbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T10:53:28Z\\\",\\\"message\\\":\\\"d as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.480341 6812 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 10:53:28.479155 6812 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1002 10:53:28.480750 6812 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 10:53:28.480839 6812 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:53:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgl42\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pccq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.802235 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9aefc404-7ddd-42cd-9ad0-1ee4004b0db2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ed1ee9f960a371b2c99ad733b029bb97a372e98a8c4c82af13702687faa6f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f42eb208457cfd66eb0c3c9e95a66fe679852aa97f5cbbbdcb7008019d00568d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.820095 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4e95cbdf-8d66-41aa-ac44-d2fbbb953936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://847742a249df92fac65c3928505433c1257ea46330e9e037aae134dd3c96a2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc7f6417b34b3a9e191ad174baa629771736434c1d87769c7e3b28ab4af46a6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ac8e15b3fe946df3db55e1c5480bc22ab840611e96e7f7a472b506173e250f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01544f2e2d79df7e412091976dfaf6211fb2898978137fd59606a22a8a07fd20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4451b9792eb9054438c8211cba674fdc605083e80a5332f21c766f693835ab6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T10:52:24Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 10:52:19.102091 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 10:52:19.103285 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3672285761/tls.crt::/tmp/serving-cert-3672285761/tls.key\\\\\\\"\\\\nI1002 10:52:24.582352 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 10:52:24.590442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 10:52:24.590488 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 10:52:24.590536 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 10:52:24.590548 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 10:52:24.602751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 10:52:24.602800 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602809 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 10:52:24.602819 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 10:52:24.602827 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 10:52:24.602834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nI1002 10:52:24.602812 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 10:52:24.602843 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 10:52:24.605144 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ea77dde0d805bfd31440ecc359d315c028a9276b424c3100b4e5dbf455948f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c506c98d5f750698a66f6e122e23140ea984161996b23667985bece2ffa9c1b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.837976 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.838054 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.838092 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.838126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.838147 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.840507 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b01be551970b9159b30aa0181e3a788d2ae870ed5c2554c14dd3262af05470d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.856704 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zzb9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb39af81-e656-4533-8b2e-9058f44dc246\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b6fe017c9ef1e714a67def17bff4f8cf540c6268d2636f25591f5941449aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvgn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zzb9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.879765 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7h88l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdca1606-9dd6-472f-bb38-4f2ac994f820\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10e314b566c999f04f9c7f49dc7ff177d87d0bafbb3a14a0b50ab773ce581a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b1a893bbe92ecd8e83d2abb146d784b1ae85dd0f75163f2a48ba95cbc2f9c80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb3301468124585ceeb650edae56842bfa020baf89802fe6dff26335832a54e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1fa23893477d727e7bb0ba548b80e2b04d43c734335167046d2387e4d58f642\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48115af3fa99d688a35522f31637d187c478cdb53aa8e5cafa2ee63576515c33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ef73452e8d5581e579b7059e32b8ddb7d6572d07517fbcbca4cda02a6c7657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed250024cc74f26a26d1e5fcfba804ebdb9e730941741fd0dc8646820ca62c91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wng8h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7h88l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.894658 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"278152b8-cf19-4211-95c9-2a162ff9549f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8z4h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzdtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.937351 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4558a8a3-d9af-4cb8-954a-ee57ebe76370\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2fce79d0c287c78276c423a4da66591c96c894ab527adcb72fc07bd2d1e0cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://053c003e68ab876f5ce044e94c140083a0387c5e2fb40f70eb57b96551b28cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bac1f5f3060b21b391cae56b30d714b25dd6e311c123523acaf77901d8e1354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ede217fd4c602d51b6800d65fb9132b04b8ad8c0bc0e051016bae140e015321b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a29f9a94eda077e73d5b97cc2454a48ec497825258a0f78f5f4cddf7179fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4931db6f28e88eb4517280685ad8f3279bbee118f8b86d102cb23ec9214a8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://600cd1d8e2b9f10da1758aed3c0d3c5dc193220a5d4f53dacae73ed637df0803\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93a5b75c919fd08f30ecc0e44be3e48330a4d0f7301114639d36e7de530f311c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T10:52:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T10:52:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.940992 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.941058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.941083 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.941113 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.941135 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:49Z","lastTransitionTime":"2025-10-02T10:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:49 crc kubenswrapper[4751]: I1002 10:53:49.958733 4751 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cba28-f92b-4fae-9df3-d67f7f4b1b4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T10:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd7dd97e3a26d1bb393ab95a7940e39dfa2113e5e5976bb42a1b383b627c1fcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f64541a9d490ab183fbebe168e337909af3d145d6108ea26e03b334c0a4ea41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T10:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xc9zs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T10:52:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-sg8vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T10:53:49Z is after 2025-08-24T17:21:41Z" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.045245 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.045309 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.045332 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.045361 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.045382 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.148630 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.148687 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.148706 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.148728 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.148746 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.251849 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.251927 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.251955 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.251988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.252011 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.355947 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.356008 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.356032 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.356060 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.356087 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.458907 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.458973 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.458989 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.459009 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.459022 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.561588 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.561696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.561715 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.561736 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.561755 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.664453 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.664508 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.664525 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.664548 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.664565 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.767167 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.767369 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.767390 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.767414 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.767435 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.870609 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.870681 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.870723 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.870755 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.870776 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.973028 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.973090 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.973109 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.973131 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:50 crc kubenswrapper[4751]: I1002 10:53:50.973147 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:50Z","lastTransitionTime":"2025-10-02T10:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.076610 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.076688 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.076713 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.076745 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.076769 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.180101 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.180210 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.180240 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.180270 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.180315 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.282881 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.282960 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.282984 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.283012 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.283037 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.385593 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.385656 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.385673 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.385696 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.385713 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.489367 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.489442 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.489465 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.489493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.489516 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.549339 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.549373 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.549463 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:51 crc kubenswrapper[4751]: E1002 10:53:51.549660 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.549711 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:51 crc kubenswrapper[4751]: E1002 10:53:51.549837 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:51 crc kubenswrapper[4751]: E1002 10:53:51.549993 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:51 crc kubenswrapper[4751]: E1002 10:53:51.550132 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.592323 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.592396 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.592421 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.592450 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.592473 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.696211 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.696271 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.696288 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.696313 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.696332 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.800284 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.800344 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.800357 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.800382 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.800408 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.903642 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.903739 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.903761 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.903788 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:51 crc kubenswrapper[4751]: I1002 10:53:51.903807 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:51Z","lastTransitionTime":"2025-10-02T10:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.006486 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.006566 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.006581 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.006600 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.006615 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.109651 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.109716 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.109732 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.109751 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.109763 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.212859 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.212913 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.212931 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.212954 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.212970 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.315126 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.315217 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.315241 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.315269 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.315290 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.418272 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.418339 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.418358 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.418378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.418393 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.520622 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.520667 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.520677 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.520695 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.520704 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.623596 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.623647 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.623660 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.623681 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.623694 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.727022 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.727102 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.727119 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.727146 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.727164 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.830378 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.830450 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.830467 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.830491 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.830509 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.933505 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.933598 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.933615 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.933640 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:52 crc kubenswrapper[4751]: I1002 10:53:52.933689 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:52Z","lastTransitionTime":"2025-10-02T10:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.036860 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.036938 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.036952 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.036975 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.036989 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.140058 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.140152 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.140224 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.140263 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.140289 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.242962 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.243026 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.243045 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.243070 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.243088 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.347016 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.347097 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.347123 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.347153 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.347209 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.450333 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.450429 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.450458 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.450481 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.450498 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.550041 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.550393 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.550477 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:53 crc kubenswrapper[4751]: E1002 10:53:53.550624 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.550969 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:53 crc kubenswrapper[4751]: E1002 10:53:53.551115 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:53 crc kubenswrapper[4751]: E1002 10:53:53.551504 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:53 crc kubenswrapper[4751]: E1002 10:53:53.551628 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.553120 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.553215 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.553238 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.553267 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.553330 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.657055 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.657111 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.657124 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.657147 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.657159 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.760346 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.760396 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.760408 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.760428 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.760441 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.863885 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.863978 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.863999 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.864025 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.864043 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.966900 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.966988 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.967013 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.967046 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:53 crc kubenswrapper[4751]: I1002 10:53:53.967072 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:53Z","lastTransitionTime":"2025-10-02T10:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.070506 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.070574 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.070592 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.070617 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.070636 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:54Z","lastTransitionTime":"2025-10-02T10:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.173814 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.173863 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.173876 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.173895 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.173912 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:54Z","lastTransitionTime":"2025-10-02T10:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.267441 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.267493 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.267501 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.267517 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.267526 4751 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T10:53:54Z","lastTransitionTime":"2025-10-02T10:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.335410 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt"] Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.336103 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.339033 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.339296 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.340020 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.342029 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.409754 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gvqc8" podStartSLOduration=84.409734304 podStartE2EDuration="1m24.409734304s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.409585419 +0000 UTC m=+116.463811899" watchObservedRunningTime="2025-10-02 10:53:54.409734304 +0000 UTC m=+116.463960764" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.429394 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.429525 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.429564 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02b85aff-c180-497b-b017-35c7f1f58395-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.429612 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02b85aff-c180-497b-b017-35c7f1f58395-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.429676 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02b85aff-c180-497b-b017-35c7f1f58395-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.453231 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8vj4p" podStartSLOduration=84.453215133 podStartE2EDuration="1m24.453215133s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.453142831 +0000 UTC m=+116.507369281" watchObservedRunningTime="2025-10-02 10:53:54.453215133 +0000 UTC m=+116.507441583" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.466037 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=60.466012541 podStartE2EDuration="1m0.466012541s" podCreationTimestamp="2025-10-02 10:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.46597898 +0000 UTC m=+116.520205430" watchObservedRunningTime="2025-10-02 10:53:54.466012541 +0000 UTC m=+116.520239031" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.499706 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.499686644 podStartE2EDuration="1m30.499686644s" podCreationTimestamp="2025-10-02 10:52:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.484753659 +0000 UTC m=+116.538980119" watchObservedRunningTime="2025-10-02 10:53:54.499686644 +0000 UTC m=+116.553913094" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.511374 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zzb9w" podStartSLOduration=84.511358301 podStartE2EDuration="1m24.511358301s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.511062383 +0000 UTC m=+116.565288833" watchObservedRunningTime="2025-10-02 10:53:54.511358301 +0000 UTC m=+116.565584751" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531045 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531094 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02b85aff-c180-497b-b017-35c7f1f58395-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531133 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02b85aff-c180-497b-b017-35c7f1f58395-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02b85aff-c180-497b-b017-35c7f1f58395-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531233 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531302 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.531347 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/02b85aff-c180-497b-b017-35c7f1f58395-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.532323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02b85aff-c180-497b-b017-35c7f1f58395-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.539285 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-7h88l" podStartSLOduration=84.539260578 podStartE2EDuration="1m24.539260578s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.539068133 +0000 UTC m=+116.593294583" watchObservedRunningTime="2025-10-02 10:53:54.539260578 +0000 UTC m=+116.593487038" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.540296 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02b85aff-c180-497b-b017-35c7f1f58395-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.550775 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:53:54 crc kubenswrapper[4751]: E1002 10:53:54.551059 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.562427 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02b85aff-c180-497b-b017-35c7f1f58395-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hkggt\" (UID: \"02b85aff-c180-497b-b017-35c7f1f58395\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.568069 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=26.568044389 podStartE2EDuration="26.568044389s" podCreationTimestamp="2025-10-02 10:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.566800455 +0000 UTC m=+116.621026945" watchObservedRunningTime="2025-10-02 10:53:54.568044389 +0000 UTC m=+116.622270889" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.584650 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-sg8vz" podStartSLOduration=83.584626859 podStartE2EDuration="1m23.584626859s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.583734815 +0000 UTC m=+116.637961305" watchObservedRunningTime="2025-10-02 10:53:54.584626859 +0000 UTC m=+116.638853319" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.623486 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=23.623458863 podStartE2EDuration="23.623458863s" podCreationTimestamp="2025-10-02 10:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.623324549 +0000 UTC m=+116.677551029" watchObservedRunningTime="2025-10-02 10:53:54.623458863 +0000 UTC m=+116.677685303" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.653232 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" Oct 02 10:53:54 crc kubenswrapper[4751]: I1002 10:53:54.680235 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podStartSLOduration=84.680210803 podStartE2EDuration="1m24.680210803s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.663858589 +0000 UTC m=+116.718085049" watchObservedRunningTime="2025-10-02 10:53:54.680210803 +0000 UTC m=+116.734437253" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.213093 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" event={"ID":"02b85aff-c180-497b-b017-35c7f1f58395","Type":"ContainerStarted","Data":"be98940c3f465bf9a50a41d4e8c6bf6c31e52138e0c8d7ebc99c6ce737caec32"} Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.213156 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" event={"ID":"02b85aff-c180-497b-b017-35c7f1f58395","Type":"ContainerStarted","Data":"c2e3445ffc45ffb07908f7fc99dff6c5f71cb95b04021d4f390c42916c1f60c1"} Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.233833 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hkggt" podStartSLOduration=85.233817804 podStartE2EDuration="1m25.233817804s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:55.233490106 +0000 UTC m=+117.287716566" watchObservedRunningTime="2025-10-02 10:53:55.233817804 +0000 UTC m=+117.288044264" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.234218 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=84.234209375 podStartE2EDuration="1m24.234209375s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:53:54.679239456 +0000 UTC m=+116.733465916" watchObservedRunningTime="2025-10-02 10:53:55.234209375 +0000 UTC m=+117.288435835" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.549653 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.549685 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:55 crc kubenswrapper[4751]: E1002 10:53:55.549780 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.549841 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:55 crc kubenswrapper[4751]: I1002 10:53:55.549925 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:55 crc kubenswrapper[4751]: E1002 10:53:55.549885 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:55 crc kubenswrapper[4751]: E1002 10:53:55.550052 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:55 crc kubenswrapper[4751]: E1002 10:53:55.550276 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:57 crc kubenswrapper[4751]: I1002 10:53:57.549494 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:57 crc kubenswrapper[4751]: I1002 10:53:57.549547 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:57 crc kubenswrapper[4751]: I1002 10:53:57.549578 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:57 crc kubenswrapper[4751]: E1002 10:53:57.549833 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:57 crc kubenswrapper[4751]: I1002 10:53:57.549877 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:57 crc kubenswrapper[4751]: E1002 10:53:57.550024 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:57 crc kubenswrapper[4751]: E1002 10:53:57.550234 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:57 crc kubenswrapper[4751]: E1002 10:53:57.550436 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.456954 4751 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 10:53:59 crc kubenswrapper[4751]: I1002 10:53:59.550112 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.551347 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:53:59 crc kubenswrapper[4751]: I1002 10:53:59.551574 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:53:59 crc kubenswrapper[4751]: I1002 10:53:59.551616 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.551674 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:53:59 crc kubenswrapper[4751]: I1002 10:53:59.551797 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.551860 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.552018 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:53:59 crc kubenswrapper[4751]: E1002 10:53:59.908456 4751 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:54:01 crc kubenswrapper[4751]: I1002 10:54:01.550213 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:01 crc kubenswrapper[4751]: I1002 10:54:01.550269 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:01 crc kubenswrapper[4751]: I1002 10:54:01.550413 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:01 crc kubenswrapper[4751]: I1002 10:54:01.550462 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:01 crc kubenswrapper[4751]: E1002 10:54:01.550630 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:01 crc kubenswrapper[4751]: E1002 10:54:01.550871 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:01 crc kubenswrapper[4751]: E1002 10:54:01.551091 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:01 crc kubenswrapper[4751]: E1002 10:54:01.551197 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:03 crc kubenswrapper[4751]: I1002 10:54:03.549719 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:03 crc kubenswrapper[4751]: I1002 10:54:03.549822 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:03 crc kubenswrapper[4751]: I1002 10:54:03.549862 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:03 crc kubenswrapper[4751]: E1002 10:54:03.550051 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:03 crc kubenswrapper[4751]: I1002 10:54:03.550118 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:03 crc kubenswrapper[4751]: E1002 10:54:03.550315 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:03 crc kubenswrapper[4751]: E1002 10:54:03.550481 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:03 crc kubenswrapper[4751]: E1002 10:54:03.550638 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:04 crc kubenswrapper[4751]: E1002 10:54:04.909917 4751 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:54:05 crc kubenswrapper[4751]: I1002 10:54:05.549776 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:05 crc kubenswrapper[4751]: I1002 10:54:05.549854 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:05 crc kubenswrapper[4751]: E1002 10:54:05.549957 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:05 crc kubenswrapper[4751]: I1002 10:54:05.550033 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:05 crc kubenswrapper[4751]: I1002 10:54:05.550084 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:05 crc kubenswrapper[4751]: E1002 10:54:05.550216 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:05 crc kubenswrapper[4751]: E1002 10:54:05.550318 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:05 crc kubenswrapper[4751]: E1002 10:54:05.550396 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.253403 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/1.log" Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.254685 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/0.log" Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.254755 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e66e701-8712-4399-b2a8-b42e8cf716a3" containerID="ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2" exitCode=1 Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.254810 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerDied","Data":"ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2"} Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.254879 4751 scope.go:117] "RemoveContainer" containerID="6e132fd18c4bd8265c885cdedfa20b418be30fe672f6400a1633ccbaaa0a88e8" Oct 02 10:54:06 crc kubenswrapper[4751]: I1002 10:54:06.255588 4751 scope.go:117] "RemoveContainer" containerID="ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2" Oct 02 10:54:06 crc kubenswrapper[4751]: E1002 10:54:06.255922 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8vj4p_openshift-multus(5e66e701-8712-4399-b2a8-b42e8cf716a3)\"" pod="openshift-multus/multus-8vj4p" podUID="5e66e701-8712-4399-b2a8-b42e8cf716a3" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.261266 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/1.log" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.549702 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:07 crc kubenswrapper[4751]: E1002 10:54:07.550224 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.550677 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:07 crc kubenswrapper[4751]: E1002 10:54:07.550922 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.551298 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:07 crc kubenswrapper[4751]: E1002 10:54:07.551566 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.551341 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:07 crc kubenswrapper[4751]: E1002 10:54:07.552248 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:07 crc kubenswrapper[4751]: I1002 10:54:07.552730 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:54:07 crc kubenswrapper[4751]: E1002 10:54:07.553052 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pccq9_openshift-ovn-kubernetes(37ba8a11-317d-406b-8628-fcd43b164dbd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" Oct 02 10:54:09 crc kubenswrapper[4751]: I1002 10:54:09.549377 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:09 crc kubenswrapper[4751]: I1002 10:54:09.549420 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:09 crc kubenswrapper[4751]: I1002 10:54:09.551824 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:09 crc kubenswrapper[4751]: E1002 10:54:09.551875 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:09 crc kubenswrapper[4751]: I1002 10:54:09.551844 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:09 crc kubenswrapper[4751]: E1002 10:54:09.551974 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:09 crc kubenswrapper[4751]: E1002 10:54:09.552082 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:09 crc kubenswrapper[4751]: E1002 10:54:09.552264 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:09 crc kubenswrapper[4751]: E1002 10:54:09.910929 4751 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:54:11 crc kubenswrapper[4751]: I1002 10:54:11.549442 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:11 crc kubenswrapper[4751]: I1002 10:54:11.549601 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:11 crc kubenswrapper[4751]: I1002 10:54:11.549818 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:11 crc kubenswrapper[4751]: E1002 10:54:11.549682 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:11 crc kubenswrapper[4751]: E1002 10:54:11.549905 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:11 crc kubenswrapper[4751]: I1002 10:54:11.549713 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:11 crc kubenswrapper[4751]: E1002 10:54:11.550000 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:11 crc kubenswrapper[4751]: E1002 10:54:11.550101 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:13 crc kubenswrapper[4751]: I1002 10:54:13.549905 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:13 crc kubenswrapper[4751]: I1002 10:54:13.549940 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:13 crc kubenswrapper[4751]: I1002 10:54:13.550091 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:13 crc kubenswrapper[4751]: E1002 10:54:13.550326 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:13 crc kubenswrapper[4751]: I1002 10:54:13.550354 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:13 crc kubenswrapper[4751]: E1002 10:54:13.550481 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:13 crc kubenswrapper[4751]: E1002 10:54:13.550647 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:13 crc kubenswrapper[4751]: E1002 10:54:13.550784 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:14 crc kubenswrapper[4751]: E1002 10:54:14.912701 4751 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:54:15 crc kubenswrapper[4751]: I1002 10:54:15.549379 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:15 crc kubenswrapper[4751]: I1002 10:54:15.549415 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:15 crc kubenswrapper[4751]: I1002 10:54:15.549436 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:15 crc kubenswrapper[4751]: I1002 10:54:15.549396 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:15 crc kubenswrapper[4751]: E1002 10:54:15.549555 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:15 crc kubenswrapper[4751]: E1002 10:54:15.549762 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:15 crc kubenswrapper[4751]: E1002 10:54:15.549849 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:15 crc kubenswrapper[4751]: E1002 10:54:15.550126 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:17 crc kubenswrapper[4751]: I1002 10:54:17.549584 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:17 crc kubenswrapper[4751]: I1002 10:54:17.549798 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:17 crc kubenswrapper[4751]: E1002 10:54:17.549875 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:17 crc kubenswrapper[4751]: I1002 10:54:17.549936 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:17 crc kubenswrapper[4751]: I1002 10:54:17.550029 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:17 crc kubenswrapper[4751]: E1002 10:54:17.550123 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:17 crc kubenswrapper[4751]: E1002 10:54:17.550306 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:17 crc kubenswrapper[4751]: E1002 10:54:17.550515 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:17 crc kubenswrapper[4751]: I1002 10:54:17.551337 4751 scope.go:117] "RemoveContainer" containerID="ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2" Oct 02 10:54:18 crc kubenswrapper[4751]: I1002 10:54:18.308718 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/1.log" Oct 02 10:54:18 crc kubenswrapper[4751]: I1002 10:54:18.308809 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerStarted","Data":"4c0bf2e16e4a69962d135faba19b1da383c6457fc10f7ded584b28ab1d7ab51a"} Oct 02 10:54:19 crc kubenswrapper[4751]: I1002 10:54:19.549461 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:19 crc kubenswrapper[4751]: I1002 10:54:19.549575 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:19 crc kubenswrapper[4751]: E1002 10:54:19.549704 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:19 crc kubenswrapper[4751]: I1002 10:54:19.549821 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:19 crc kubenswrapper[4751]: I1002 10:54:19.549888 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:19 crc kubenswrapper[4751]: E1002 10:54:19.552030 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:19 crc kubenswrapper[4751]: E1002 10:54:19.552234 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:19 crc kubenswrapper[4751]: E1002 10:54:19.552289 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:19 crc kubenswrapper[4751]: E1002 10:54:19.913444 4751 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:54:20 crc kubenswrapper[4751]: I1002 10:54:20.550537 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.320797 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/3.log" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.323577 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerStarted","Data":"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3"} Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.324668 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.364025 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podStartSLOduration=110.364011303 podStartE2EDuration="1m50.364011303s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:21.360952727 +0000 UTC m=+143.415179187" watchObservedRunningTime="2025-10-02 10:54:21.364011303 +0000 UTC m=+143.418237753" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.549931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.549986 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.549931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.549929 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:21 crc kubenswrapper[4751]: E1002 10:54:21.550063 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:21 crc kubenswrapper[4751]: E1002 10:54:21.550146 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:21 crc kubenswrapper[4751]: E1002 10:54:21.550316 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:21 crc kubenswrapper[4751]: E1002 10:54:21.550471 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:21 crc kubenswrapper[4751]: I1002 10:54:21.726288 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzdtd"] Oct 02 10:54:22 crc kubenswrapper[4751]: I1002 10:54:22.327687 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:22 crc kubenswrapper[4751]: E1002 10:54:22.328492 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:23 crc kubenswrapper[4751]: I1002 10:54:23.549991 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:23 crc kubenswrapper[4751]: I1002 10:54:23.550037 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:23 crc kubenswrapper[4751]: I1002 10:54:23.550037 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:23 crc kubenswrapper[4751]: E1002 10:54:23.550153 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 10:54:23 crc kubenswrapper[4751]: E1002 10:54:23.550403 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 10:54:23 crc kubenswrapper[4751]: E1002 10:54:23.550558 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 10:54:24 crc kubenswrapper[4751]: I1002 10:54:24.549115 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:24 crc kubenswrapper[4751]: E1002 10:54:24.549380 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzdtd" podUID="278152b8-cf19-4211-95c9-2a162ff9549f" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.549974 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.550007 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.550234 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.552644 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.552653 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.553273 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 10:54:25 crc kubenswrapper[4751]: I1002 10:54:25.554063 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 10:54:26 crc kubenswrapper[4751]: I1002 10:54:26.549376 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:26 crc kubenswrapper[4751]: I1002 10:54:26.552371 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 10:54:26 crc kubenswrapper[4751]: I1002 10:54:26.553419 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 10:54:31 crc kubenswrapper[4751]: I1002 10:54:31.507332 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:54:31 crc kubenswrapper[4751]: I1002 10:54:31.507440 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:54:31 crc kubenswrapper[4751]: I1002 10:54:31.947091 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.287352 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.287597 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:33 crc kubenswrapper[4751]: E1002 10:54:33.287661 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:56:35.287613416 +0000 UTC m=+277.341840006 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.287771 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.289032 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.298007 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.369442 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.389575 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.389667 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.395673 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.395690 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:33 crc kubenswrapper[4751]: W1002 10:54:33.610303 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-1b4394da10bccfb69ebb5d52f684b436d2aa6d25025ce59f2a7ca133fd9b82a2 WatchSource:0}: Error finding container 1b4394da10bccfb69ebb5d52f684b436d2aa6d25025ce59f2a7ca133fd9b82a2: Status 404 returned error can't find the container with id 1b4394da10bccfb69ebb5d52f684b436d2aa6d25025ce59f2a7ca133fd9b82a2 Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.678908 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 10:54:33 crc kubenswrapper[4751]: I1002 10:54:33.692161 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:33 crc kubenswrapper[4751]: W1002 10:54:33.873281 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-277f0e6623b5ccf54720dfaf07b575d0c3390e6934e56c9d13413bac9c76e8ad WatchSource:0}: Error finding container 277f0e6623b5ccf54720dfaf07b575d0c3390e6934e56c9d13413bac9c76e8ad: Status 404 returned error can't find the container with id 277f0e6623b5ccf54720dfaf07b575d0c3390e6934e56c9d13413bac9c76e8ad Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.373338 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7a5db9e42bc7449be763cd429425f174b4fbea8d7da75a8876b8ef93c6fc04db"} Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.373433 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0a46edf8d695bec0dfaf977ad238b8d23b736a072add8559f44d84e968e8f2c7"} Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.373690 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.376603 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d15be0056ff7ea49e6cbde851c0cf55e399b4824b4ebff02e3f2a6a6b09b6a45"} Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.376810 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"277f0e6623b5ccf54720dfaf07b575d0c3390e6934e56c9d13413bac9c76e8ad"} Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.378961 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8ca511ba77d18566e55812ecdc2e04dd99439fb18d1e7ecd410e3c08d40ce3e2"} Oct 02 10:54:34 crc kubenswrapper[4751]: I1002 10:54:34.379040 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1b4394da10bccfb69ebb5d52f684b436d2aa6d25025ce59f2a7ca133fd9b82a2"} Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.079708 4751 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.134284 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.135097 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.136530 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.137418 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.138220 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.139241 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.139955 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.140905 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.142285 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6tkns"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.143345 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.143376 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.148579 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.149588 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.150264 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gg8df"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.151105 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.160475 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.161403 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.165550 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.166817 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.168400 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.168512 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.181951 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.186160 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.186598 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.187029 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.187108 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.190010 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.190334 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.190510 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.190773 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.190964 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.191262 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.191465 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.191677 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.191863 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192021 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192246 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192425 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192477 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192542 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192650 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192686 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192429 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.192862 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194009 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194145 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194464 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194601 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194685 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194793 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194898 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.194929 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.195002 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.195113 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.195260 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.195370 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.195460 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.196943 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.197043 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.197123 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.197437 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.197746 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.197865 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.199050 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mrvdc"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.199484 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.199703 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7lhjm"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.199980 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.200284 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.200508 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.203217 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.203317 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.205632 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.205750 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.207226 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.207791 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2t6m9"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.208123 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4r69k"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.208634 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.209079 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.209434 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.211679 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.212826 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.217218 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.228915 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.230701 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.231062 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.231234 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.231495 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.231893 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.232132 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.231549 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.232779 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.232598 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.235964 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.236362 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.236648 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.236886 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.237562 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.237594 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.237957 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.238263 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.238699 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.239359 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.239640 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.239813 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.240732 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.249723 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.251712 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2dt22"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.252521 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.253153 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.254382 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.253696 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.253740 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.254645 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.254761 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.256631 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.256704 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.256966 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.257151 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.257424 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.257798 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.258383 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.257805 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.259159 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.261370 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.261549 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.262129 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.264630 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.265091 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.267511 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.269424 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.269567 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.269695 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.269822 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.269960 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.277802 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.278102 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.278359 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.282905 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.283402 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.283825 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.284199 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.284485 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.284643 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.285254 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.286002 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.286039 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.286761 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.286832 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.292099 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.293329 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jnthj"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.293787 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.293859 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.294527 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.298899 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.299226 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.299770 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.299857 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.302455 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ks9q4"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.303846 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.304163 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.304277 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.304555 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.306242 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.306288 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.306784 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.306938 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307548 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-config\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307579 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a6cb79-0b3a-49b0-9b5c-580030f9978d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307609 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkbf\" (UniqueName: \"kubernetes.io/projected/40a6cb79-0b3a-49b0-9b5c-580030f9978d-kube-api-access-cqkbf\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307627 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mcwc\" (UniqueName: \"kubernetes.io/projected/5fcfe397-0564-4c53-b2c2-0b1bc0796269-kube-api-access-7mcwc\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307648 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5057ef7-7363-4d5c-b72c-d7473c37696d-serving-cert\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307695 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307714 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-service-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307721 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307730 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307761 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fcfe397-0564-4c53-b2c2-0b1bc0796269-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307785 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307806 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-audit-dir\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307834 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307860 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307876 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-serving-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307897 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xlnq\" (UniqueName: \"kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307913 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307940 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp852\" (UniqueName: \"kubernetes.io/projected/f33d078a-62a0-4f6e-8efb-85003d873289-kube-api-access-rp852\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.307974 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-policies\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308085 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a6cb79-0b3a-49b0-9b5c-580030f9978d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308110 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-dir\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308135 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308155 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308194 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-encryption-config\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308465 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308505 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae90eabd-ce14-4792-b171-589323678ab0-metrics-tls\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308533 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308576 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-node-pullsecrets\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308596 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-audit\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308631 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308657 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-client\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308689 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-client\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308747 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rblmg\" (UniqueName: \"kubernetes.io/projected/d5057ef7-7363-4d5c-b72c-d7473c37696d-kube-api-access-rblmg\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308774 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308805 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308822 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33d078a-62a0-4f6e-8efb-85003d873289-serving-cert\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308865 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308882 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfwc4\" (UniqueName: \"kubernetes.io/projected/176d9f58-b454-4171-8297-f295b8650956-kube-api-access-wfwc4\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308906 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.308946 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f33d078a-62a0-4f6e-8efb-85003d873289-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309004 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309032 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309054 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g2j6\" (UniqueName: \"kubernetes.io/projected/ae90eabd-ce14-4792-b171-589323678ab0-kube-api-access-5g2j6\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309071 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309427 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-image-import-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309463 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309491 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.309516 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-config\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.310655 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mrvdc"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.310704 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.311674 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.311856 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-serving-cert\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.311902 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.311952 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.311990 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cp5v\" (UniqueName: \"kubernetes.io/projected/2357ddd4-8799-463c-82e6-e98fe53e328b-kube-api-access-9cp5v\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312011 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/176d9f58-b454-4171-8297-f295b8650956-serving-cert\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312043 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-encryption-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312065 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-trusted-ca\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312088 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312145 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsw2b\" (UniqueName: \"kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312255 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312276 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq27z\" (UniqueName: \"kubernetes.io/projected/763d7fee-c2ed-4f24-8ff3-27f308054771-kube-api-access-rq27z\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312419 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrt97\" (UniqueName: \"kubernetes.io/projected/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-kube-api-access-wrt97\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312481 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-serving-cert\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312518 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.312783 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktsq\" (UniqueName: \"kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.317084 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.319866 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.320824 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.321756 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.330320 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.331702 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.339037 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rr827"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.345001 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.351666 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.351833 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.353513 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.354055 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.355081 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.355892 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6tkns"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.357525 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.358362 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.359766 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.361068 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.361858 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-swmnp"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.362633 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.363714 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-v4ns6"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.365149 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.369100 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2t6m9"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.370576 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7lhjm"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.372408 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.374320 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.376862 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.377375 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.377496 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.380644 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.383535 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.385489 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4r69k"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.385524 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.387984 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.390408 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.397518 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.396967 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.401554 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.402844 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gg8df"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.403884 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2dt22"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.404935 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7q7g2"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.405623 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.406490 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bwnql"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.407612 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.407845 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.408750 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.410702 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.411656 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413386 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413415 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-serving-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413434 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xlnq\" (UniqueName: \"kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413453 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413474 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720be237-ff1f-4ed2-b55b-6e46c976cce8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413497 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413514 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp852\" (UniqueName: \"kubernetes.io/projected/f33d078a-62a0-4f6e-8efb-85003d873289-kube-api-access-rp852\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413532 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-policies\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413552 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a6cb79-0b3a-49b0-9b5c-580030f9978d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413589 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413606 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413622 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-encryption-config\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413637 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-dir\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413654 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413682 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkrwn\" (UniqueName: \"kubernetes.io/projected/58b8cdd6-dfa9-4e52-8f93-12663a522318-kube-api-access-xkrwn\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413702 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720be237-ff1f-4ed2-b55b-6e46c976cce8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413719 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae90eabd-ce14-4792-b171-589323678ab0-metrics-tls\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413737 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413756 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-node-pullsecrets\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413773 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-audit\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413792 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413810 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413828 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-client\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413845 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-images\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413862 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-client\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413880 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0413ee62-e2ef-4ca0-a21f-2990f11169c3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413900 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413917 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rblmg\" (UniqueName: \"kubernetes.io/projected/d5057ef7-7363-4d5c-b72c-d7473c37696d-kube-api-access-rblmg\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413936 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413954 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413972 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33d078a-62a0-4f6e-8efb-85003d873289-serving-cert\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.413989 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414005 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfwc4\" (UniqueName: \"kubernetes.io/projected/176d9f58-b454-4171-8297-f295b8650956-kube-api-access-wfwc4\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414021 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f33d078a-62a0-4f6e-8efb-85003d873289-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414060 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414076 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414092 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414109 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414127 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g2j6\" (UniqueName: \"kubernetes.io/projected/ae90eabd-ce14-4792-b171-589323678ab0-kube-api-access-5g2j6\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414141 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-image-import-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414156 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414215 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-config\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414234 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndbdq\" (UniqueName: \"kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414252 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-serving-cert\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414268 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414284 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414317 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cp5v\" (UniqueName: \"kubernetes.io/projected/2357ddd4-8799-463c-82e6-e98fe53e328b-kube-api-access-9cp5v\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414336 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/176d9f58-b454-4171-8297-f295b8650956-serving-cert\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414358 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-encryption-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414379 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-trusted-ca\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414400 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414419 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414436 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsw2b\" (UniqueName: \"kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414497 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq27z\" (UniqueName: \"kubernetes.io/projected/763d7fee-c2ed-4f24-8ff3-27f308054771-kube-api-access-rq27z\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414515 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720be237-ff1f-4ed2-b55b-6e46c976cce8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414532 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrt97\" (UniqueName: \"kubernetes.io/projected/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-kube-api-access-wrt97\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414548 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-serving-cert\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414563 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414587 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414605 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktsq\" (UniqueName: \"kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414621 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-config\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414636 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a6cb79-0b3a-49b0-9b5c-580030f9978d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414651 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkbf\" (UniqueName: \"kubernetes.io/projected/40a6cb79-0b3a-49b0-9b5c-580030f9978d-kube-api-access-cqkbf\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414667 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mcwc\" (UniqueName: \"kubernetes.io/projected/5fcfe397-0564-4c53-b2c2-0b1bc0796269-kube-api-access-7mcwc\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414683 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5057ef7-7363-4d5c-b72c-d7473c37696d-serving-cert\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414700 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414726 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414741 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-service-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414758 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfsgp\" (UniqueName: \"kubernetes.io/projected/0413ee62-e2ef-4ca0-a21f-2990f11169c3-kube-api-access-dfsgp\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414779 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414794 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fcfe397-0564-4c53-b2c2-0b1bc0796269-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414814 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414837 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-audit-dir\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414861 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.414883 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58b8cdd6-dfa9-4e52-8f93-12663a522318-proxy-tls\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.415655 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.416664 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-image-import-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.416937 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.417046 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-node-pullsecrets\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.417636 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.417776 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.418479 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.418690 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.419641 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f33d078a-62a0-4f6e-8efb-85003d873289-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.420312 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.420359 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.420483 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-serving-ca\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.422113 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.422192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.422278 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-audit\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.422274 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.422418 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.424720 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-config\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.425659 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-policies\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.429183 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-encryption-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.429745 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-trusted-ca\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.429853 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.429845 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.430390 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33d078a-62a0-4f6e-8efb-85003d873289-serving-cert\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431011 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-serving-cert\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431102 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ks9q4"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431152 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-swmnp"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431192 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431433 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-etcd-client\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431662 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5057ef7-7363-4d5c-b72c-d7473c37696d-config\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431820 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.431964 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.432473 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40a6cb79-0b3a-49b0-9b5c-580030f9978d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.432738 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.432786 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/763d7fee-c2ed-4f24-8ff3-27f308054771-serving-cert\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.433748 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-service-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.433783 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.434438 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/763d7fee-c2ed-4f24-8ff3-27f308054771-config\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.434578 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.434938 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/763d7fee-c2ed-4f24-8ff3-27f308054771-audit-dir\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.435501 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2357ddd4-8799-463c-82e6-e98fe53e328b-audit-dir\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.437276 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/176d9f58-b454-4171-8297-f295b8650956-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.437356 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.437462 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5057ef7-7363-4d5c-b72c-d7473c37696d-serving-cert\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.437818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.439351 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.440051 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.440628 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.441126 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.442442 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/176d9f58-b454-4171-8297-f295b8650956-serving-cert\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.443824 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.444462 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.444702 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.445133 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.445372 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-encryption-config\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.445718 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae90eabd-ce14-4792-b171-589323678ab0-metrics-tls\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.445220 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.446135 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.446364 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40a6cb79-0b3a-49b0-9b5c-580030f9978d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.446454 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2357ddd4-8799-463c-82e6-e98fe53e328b-etcd-client\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.446864 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.448559 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fcfe397-0564-4c53-b2c2-0b1bc0796269-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.450721 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.451094 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.457451 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7q7g2"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.458680 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rr827"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.459130 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.463644 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bwnql"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.467451 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6swj5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.469147 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.472242 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6swj5"] Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.478304 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.498122 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515537 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkrwn\" (UniqueName: \"kubernetes.io/projected/58b8cdd6-dfa9-4e52-8f93-12663a522318-kube-api-access-xkrwn\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515557 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720be237-ff1f-4ed2-b55b-6e46c976cce8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515577 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-images\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515595 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0413ee62-e2ef-4ca0-a21f-2990f11169c3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515634 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515657 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndbdq\" (UniqueName: \"kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515676 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515718 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720be237-ff1f-4ed2-b55b-6e46c976cce8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515740 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515791 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfsgp\" (UniqueName: \"kubernetes.io/projected/0413ee62-e2ef-4ca0-a21f-2990f11169c3-kube-api-access-dfsgp\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515810 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58b8cdd6-dfa9-4e52-8f93-12663a522318-proxy-tls\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.515830 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720be237-ff1f-4ed2-b55b-6e46c976cce8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.516966 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.517593 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-auth-proxy-config\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.517617 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.518093 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.518384 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.527674 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.537397 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.550006 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/720be237-ff1f-4ed2-b55b-6e46c976cce8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.560262 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.566890 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/720be237-ff1f-4ed2-b55b-6e46c976cce8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.577126 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.598686 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.616482 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.636892 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.657000 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.677854 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.699145 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.725487 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.738885 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.758012 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.777709 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.797635 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.818436 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.838401 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.858237 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.878587 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.897198 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.919456 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.938895 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.958407 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.978639 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.987835 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/58b8cdd6-dfa9-4e52-8f93-12663a522318-images\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:35 crc kubenswrapper[4751]: I1002 10:54:35.998347 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.019068 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.033388 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58b8cdd6-dfa9-4e52-8f93-12663a522318-proxy-tls\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.039212 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.060162 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.078838 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.098820 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.118953 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.139297 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.159318 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.178092 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.197985 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.218009 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.237622 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.257868 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.276721 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.298115 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.315469 4751 request.go:700] Waited for 1.008341841s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackageserver-service-cert&limit=500&resourceVersion=0 Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.317960 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.338546 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.358483 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.371886 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0413ee62-e2ef-4ca0-a21f-2990f11169c3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.380087 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.401843 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.439300 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.467068 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.477690 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.497419 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.518014 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.539215 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.557863 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.578956 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.598790 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.617919 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.636973 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.657706 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.677614 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.698039 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.717536 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.757861 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.777931 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.797976 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.818030 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.838262 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.857684 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.877658 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.898513 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.917206 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.939160 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.959757 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.979366 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 10:54:36 crc kubenswrapper[4751]: I1002 10:54:36.998642 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.017925 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.039569 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.058027 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.078477 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.098327 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.118602 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.137355 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.184674 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g2j6\" (UniqueName: \"kubernetes.io/projected/ae90eabd-ce14-4792-b171-589323678ab0-kube-api-access-5g2j6\") pod \"dns-operator-744455d44c-7lhjm\" (UID: \"ae90eabd-ce14-4792-b171-589323678ab0\") " pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.213743 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrt97\" (UniqueName: \"kubernetes.io/projected/2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a-kube-api-access-wrt97\") pod \"openshift-apiserver-operator-796bbdcf4f-c4z9t\" (UID: \"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.227283 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xlnq\" (UniqueName: \"kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq\") pod \"console-f9d7485db-k2g7v\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.235770 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp852\" (UniqueName: \"kubernetes.io/projected/f33d078a-62a0-4f6e-8efb-85003d873289-kube-api-access-rp852\") pod \"openshift-config-operator-7777fb866f-6tkns\" (UID: \"f33d078a-62a0-4f6e-8efb-85003d873289\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.255485 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfwc4\" (UniqueName: \"kubernetes.io/projected/176d9f58-b454-4171-8297-f295b8650956-kube-api-access-wfwc4\") pod \"authentication-operator-69f744f599-2t6m9\" (UID: \"176d9f58-b454-4171-8297-f295b8650956\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.272249 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsw2b\" (UniqueName: \"kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b\") pod \"route-controller-manager-6576b87f9c-fppx5\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.287816 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.294920 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cp5v\" (UniqueName: \"kubernetes.io/projected/2357ddd4-8799-463c-82e6-e98fe53e328b-kube-api-access-9cp5v\") pod \"apiserver-7bbb656c7d-6qpgb\" (UID: \"2357ddd4-8799-463c-82e6-e98fe53e328b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.299610 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.315718 4751 request.go:700] Waited for 1.884196063s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.321256 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktsq\" (UniqueName: \"kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq\") pod \"oauth-openshift-558db77b4-ztmtr\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.330543 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq27z\" (UniqueName: \"kubernetes.io/projected/763d7fee-c2ed-4f24-8ff3-27f308054771-kube-api-access-rq27z\") pod \"apiserver-76f77b778f-gg8df\" (UID: \"763d7fee-c2ed-4f24-8ff3-27f308054771\") " pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.354402 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rblmg\" (UniqueName: \"kubernetes.io/projected/d5057ef7-7363-4d5c-b72c-d7473c37696d-kube-api-access-rblmg\") pod \"console-operator-58897d9998-mrvdc\" (UID: \"d5057ef7-7363-4d5c-b72c-d7473c37696d\") " pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.362646 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.378640 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkbf\" (UniqueName: \"kubernetes.io/projected/40a6cb79-0b3a-49b0-9b5c-580030f9978d-kube-api-access-cqkbf\") pod \"openshift-controller-manager-operator-756b6f6bc6-zzmsd\" (UID: \"40a6cb79-0b3a-49b0-9b5c-580030f9978d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.393460 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.397050 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mcwc\" (UniqueName: \"kubernetes.io/projected/5fcfe397-0564-4c53-b2c2-0b1bc0796269-kube-api-access-7mcwc\") pod \"cluster-samples-operator-665b6dd947-7t5dg\" (UID: \"5fcfe397-0564-4c53-b2c2-0b1bc0796269\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.397446 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.416865 4751 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.418919 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.426481 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.437766 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.454072 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.462434 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.477051 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.478946 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/720be237-ff1f-4ed2-b55b-6e46c976cce8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rqr28\" (UID: \"720be237-ff1f-4ed2-b55b-6e46c976cce8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.495824 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkrwn\" (UniqueName: \"kubernetes.io/projected/58b8cdd6-dfa9-4e52-8f93-12663a522318-kube-api-access-xkrwn\") pod \"machine-config-operator-74547568cd-k2xtp\" (UID: \"58b8cdd6-dfa9-4e52-8f93-12663a522318\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.498079 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.510436 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.513405 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfsgp\" (UniqueName: \"kubernetes.io/projected/0413ee62-e2ef-4ca0-a21f-2990f11169c3-kube-api-access-dfsgp\") pod \"control-plane-machine-set-operator-78cbb6b69f-p5jlr\" (UID: \"0413ee62-e2ef-4ca0-a21f-2990f11169c3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.538057 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndbdq\" (UniqueName: \"kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq\") pod \"controller-manager-879f6c89f-flnxn\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550700 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550779 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-images\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550833 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550855 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acce769-aae9-45be-b613-ff87c0d23453-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550879 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-webhook-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.550940 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed30f5ae-4c45-4928-9ef6-079587f62490-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551012 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-serving-cert\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551066 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5891fee9-8bb0-490a-8b03-7da135d49156-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551105 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8add3b3a-0164-4e25-9caf-ec05032c9b90-proxy-tls\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551125 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d278f4ff-d1f2-4710-a8f4-e87a429062a9-trusted-ca\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551144 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551187 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-auth-proxy-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551250 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acce769-aae9-45be-b613-ff87c0d23453-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551274 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551299 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97875ab4-12f1-45da-af39-353e06064013-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551323 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97875ab4-12f1-45da-af39-353e06064013-config\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551345 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-service-ca-bundle\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551379 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551399 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-service-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551434 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtvxx\" (UniqueName: \"kubernetes.io/projected/8add3b3a-0164-4e25-9caf-ec05032c9b90-kube-api-access-wtvxx\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551476 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-default-certificate\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551499 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bd01da30-84c7-4814-a15a-0ea85c1de44f-machine-approver-tls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551523 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5mv4\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-kube-api-access-p5mv4\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551543 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551590 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trmgf\" (UniqueName: \"kubernetes.io/projected/2061adbc-fb55-4f79-8d65-93729fb4afb1-kube-api-access-trmgf\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551659 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551679 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-tmpfs\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.551901 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68s6f\" (UniqueName: \"kubernetes.io/projected/f14b663f-94ad-44f2-82a8-ec6d53159146-kube-api-access-68s6f\") pod \"downloads-7954f5f757-2dt22\" (UID: \"f14b663f-94ad-44f2-82a8-ec6d53159146\") " pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.557974 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79zgg\" (UniqueName: \"kubernetes.io/projected/0acce769-aae9-45be-b613-ff87c0d23453-kube-api-access-79zgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558078 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558106 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-config\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558139 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8add3b3a-0164-4e25-9caf-ec05032c9b90-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558179 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5891fee9-8bb0-490a-8b03-7da135d49156-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558237 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d278f4ff-d1f2-4710-a8f4-e87a429062a9-metrics-tls\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558305 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbnqc\" (UniqueName: \"kubernetes.io/projected/ed30f5ae-4c45-4928-9ef6-079587f62490-kube-api-access-wbnqc\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558346 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-apiservice-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558382 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc181813-c51d-496e-ba0d-4f03f544d9d2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558406 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558505 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lbls\" (UniqueName: \"kubernetes.io/projected/bd01da30-84c7-4814-a15a-0ea85c1de44f-kube-api-access-2lbls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558545 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc181813-c51d-496e-ba0d-4f03f544d9d2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558568 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558605 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mmjw\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558628 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-metrics-certs\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2s8q\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-kube-api-access-l2s8q\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558788 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bksx\" (UniqueName: \"kubernetes.io/projected/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-kube-api-access-2bksx\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558814 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-config\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558832 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-stats-auth\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsprr\" (UniqueName: \"kubernetes.io/projected/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-kube-api-access-hsprr\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558890 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558936 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5891fee9-8bb0-490a-8b03-7da135d49156-config\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558955 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-client\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.558971 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97875ab4-12f1-45da-af39-353e06064013-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: E1002 10:54:37.565433 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.065408495 +0000 UTC m=+160.119634945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.572654 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.573165 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.601889 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.608060 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6tkns"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.616031 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" Oct 02 10:54:37 crc kubenswrapper[4751]: W1002 10:54:37.634726 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf33d078a_62a0_4f6e_8efb_85003d873289.slice/crio-53da1087eadc1a361e3ee11138ea70ab88b298b7ed8cfaa7e842a5ed210f3c90 WatchSource:0}: Error finding container 53da1087eadc1a361e3ee11138ea70ab88b298b7ed8cfaa7e842a5ed210f3c90: Status 404 returned error can't find the container with id 53da1087eadc1a361e3ee11138ea70ab88b298b7ed8cfaa7e842a5ed210f3c90 Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.636361 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.659925 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660121 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-csi-data-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660146 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc181813-c51d-496e-ba0d-4f03f544d9d2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660163 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660201 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mmjw\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660218 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-certs\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660232 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660250 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-metrics-certs\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660274 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2s8q\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-kube-api-access-l2s8q\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660290 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bksx\" (UniqueName: \"kubernetes.io/projected/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-kube-api-access-2bksx\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660307 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-config\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660322 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgszr\" (UniqueName: \"kubernetes.io/projected/9993f91a-910b-48d3-b6cd-1291630cf77b-kube-api-access-dgszr\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660346 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660360 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-stats-auth\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660374 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsprr\" (UniqueName: \"kubernetes.io/projected/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-kube-api-access-hsprr\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660390 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-srv-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660406 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5891fee9-8bb0-490a-8b03-7da135d49156-config\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660425 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-client\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660441 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97875ab4-12f1-45da-af39-353e06064013-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660480 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660497 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j8ct\" (UniqueName: \"kubernetes.io/projected/eb194ce7-018a-43ac-b414-acd2cc8e60d3-kube-api-access-4j8ct\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660510 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7cl\" (UniqueName: \"kubernetes.io/projected/cfd2757a-f362-4ecc-8396-5c8d667f1af7-kube-api-access-pd7cl\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-images\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660559 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-registration-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660614 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-config\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660635 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-node-bootstrap-token\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660657 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660679 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acce769-aae9-45be-b613-ff87c0d23453-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660698 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjr4r\" (UniqueName: \"kubernetes.io/projected/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-kube-api-access-bjr4r\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660713 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660730 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-mountpoint-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660747 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-webhook-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660766 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d33820a-6276-4e72-820e-27876200d322-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660813 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed30f5ae-4c45-4928-9ef6-079587f62490-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660831 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-metrics-tls\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660847 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w4wn\" (UniqueName: \"kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660862 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-serving-cert\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660878 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5891fee9-8bb0-490a-8b03-7da135d49156-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660896 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8add3b3a-0164-4e25-9caf-ec05032c9b90-proxy-tls\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660911 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d278f4ff-d1f2-4710-a8f4-e87a429062a9-trusted-ca\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660927 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-auth-proxy-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660944 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-socket-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660959 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660982 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-profile-collector-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.660997 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-plugins-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661011 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cfd2757a-f362-4ecc-8396-5c8d667f1af7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661024 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/345a87c1-958b-438a-b808-2b9007d74537-cert\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661038 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661057 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acce769-aae9-45be-b613-ff87c0d23453-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzhf5\" (UniqueName: \"kubernetes.io/projected/6d33820a-6276-4e72-820e-27876200d322-kube-api-access-pzhf5\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661088 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-serving-cert\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661103 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-kube-api-access-j6sht\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661119 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661143 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97875ab4-12f1-45da-af39-353e06064013-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661158 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7dsr\" (UniqueName: \"kubernetes.io/projected/e75d1cca-da41-4ef9-9fff-3bd99754c344-kube-api-access-d7dsr\") pod \"migrator-59844c95c7-dnqg5\" (UID: \"e75d1cca-da41-4ef9-9fff-3bd99754c344\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661194 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97875ab4-12f1-45da-af39-353e06064013-config\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-cabundle\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661228 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-service-ca-bundle\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661251 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-service-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtvxx\" (UniqueName: \"kubernetes.io/projected/8add3b3a-0164-4e25-9caf-ec05032c9b90-kube-api-access-wtvxx\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-default-certificate\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661345 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bd01da30-84c7-4814-a15a-0ea85c1de44f-machine-approver-tls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661370 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5mv4\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-kube-api-access-p5mv4\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661386 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661402 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trmgf\" (UniqueName: \"kubernetes.io/projected/2061adbc-fb55-4f79-8d65-93729fb4afb1-kube-api-access-trmgf\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661456 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-tmpfs\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661473 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-config-volume\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661489 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvwpl\" (UniqueName: \"kubernetes.io/projected/0868165a-d1ad-4853-8318-53034de4b53b-kube-api-access-zvwpl\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661515 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p98kp\" (UniqueName: \"kubernetes.io/projected/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-kube-api-access-p98kp\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68s6f\" (UniqueName: \"kubernetes.io/projected/f14b663f-94ad-44f2-82a8-ec6d53159146-kube-api-access-68s6f\") pod \"downloads-7954f5f757-2dt22\" (UID: \"f14b663f-94ad-44f2-82a8-ec6d53159146\") " pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661567 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k4rx\" (UniqueName: \"kubernetes.io/projected/345a87c1-958b-438a-b808-2b9007d74537-kube-api-access-4k4rx\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661590 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q442s\" (UniqueName: \"kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661632 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79zgg\" (UniqueName: \"kubernetes.io/projected/0acce769-aae9-45be-b613-ff87c0d23453-kube-api-access-79zgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661670 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-config\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661684 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-key\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661699 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661716 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8add3b3a-0164-4e25-9caf-ec05032c9b90-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661732 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-srv-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661749 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5891fee9-8bb0-490a-8b03-7da135d49156-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.661763 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d278f4ff-d1f2-4710-a8f4-e87a429062a9-metrics-tls\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662111 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbnqc\" (UniqueName: \"kubernetes.io/projected/ed30f5ae-4c45-4928-9ef6-079587f62490-kube-api-access-wbnqc\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662139 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-apiservice-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662164 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc181813-c51d-496e-ba0d-4f03f544d9d2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662216 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lbls\" (UniqueName: \"kubernetes.io/projected/bd01da30-84c7-4814-a15a-0ea85c1de44f-kube-api-access-2lbls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.662239 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4nq\" (UniqueName: \"kubernetes.io/projected/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-kube-api-access-6m4nq\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: E1002 10:54:37.664247 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.164229015 +0000 UTC m=+160.218455465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.664467 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.664613 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-auth-proxy-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.664898 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d278f4ff-d1f2-4710-a8f4-e87a429062a9-trusted-ca\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.665417 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5891fee9-8bb0-490a-8b03-7da135d49156-config\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.665526 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8add3b3a-0164-4e25-9caf-ec05032c9b90-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.666148 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-config\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.668637 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd01da30-84c7-4814-a15a-0ea85c1de44f-config\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.668806 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-tmpfs\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.669660 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.671027 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97875ab4-12f1-45da-af39-353e06064013-config\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.671470 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-images\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.671832 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bc181813-c51d-496e-ba0d-4f03f544d9d2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.671932 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-service-ca-bundle\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.671979 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d278f4ff-d1f2-4710-a8f4-e87a429062a9-metrics-tls\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.673055 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acce769-aae9-45be-b613-ff87c0d23453-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.672411 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc181813-c51d-496e-ba0d-4f03f544d9d2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.673654 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-service-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.673732 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed30f5ae-4c45-4928-9ef6-079587f62490-config\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.674043 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.674197 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.674472 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-ca\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.676111 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0acce769-aae9-45be-b613-ff87c0d23453-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.677072 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.679418 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-etcd-client\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.681021 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8add3b3a-0164-4e25-9caf-ec05032c9b90-proxy-tls\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.681086 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2061adbc-fb55-4f79-8d65-93729fb4afb1-serving-cert\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.681289 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bd01da30-84c7-4814-a15a-0ea85c1de44f-machine-approver-tls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.681522 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-metrics-certs\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.682630 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-webhook-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.682774 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-default-certificate\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.684225 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97875ab4-12f1-45da-af39-353e06064013-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.684583 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-stats-auth\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.685638 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5891fee9-8bb0-490a-8b03-7da135d49156-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.685821 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-apiservice-cert\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.687474 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed30f5ae-4c45-4928-9ef6-079587f62490-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.687482 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.691834 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbnqc\" (UniqueName: \"kubernetes.io/projected/ed30f5ae-4c45-4928-9ef6-079587f62490-kube-api-access-wbnqc\") pod \"machine-api-operator-5694c8668f-4r69k\" (UID: \"ed30f5ae-4c45-4928-9ef6-079587f62490\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.694719 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.709790 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.714844 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.727917 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.735672 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lbls\" (UniqueName: \"kubernetes.io/projected/bd01da30-84c7-4814-a15a-0ea85c1de44f-kube-api-access-2lbls\") pod \"machine-approver-56656f9798-7pnb7\" (UID: \"bd01da30-84c7-4814-a15a-0ea85c1de44f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764012 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764351 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgszr\" (UniqueName: \"kubernetes.io/projected/9993f91a-910b-48d3-b6cd-1291630cf77b-kube-api-access-dgszr\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764402 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-srv-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764447 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j8ct\" (UniqueName: \"kubernetes.io/projected/eb194ce7-018a-43ac-b414-acd2cc8e60d3-kube-api-access-4j8ct\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764480 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7cl\" (UniqueName: \"kubernetes.io/projected/cfd2757a-f362-4ecc-8396-5c8d667f1af7-kube-api-access-pd7cl\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764502 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764531 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-registration-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764556 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-config\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764581 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-node-bootstrap-token\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764603 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjr4r\" (UniqueName: \"kubernetes.io/projected/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-kube-api-access-bjr4r\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764623 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-mountpoint-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764661 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d33820a-6276-4e72-820e-27876200d322-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764684 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-metrics-tls\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764700 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w4wn\" (UniqueName: \"kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764722 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-socket-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764739 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-profile-collector-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-plugins-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764769 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cfd2757a-f362-4ecc-8396-5c8d667f1af7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764786 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/345a87c1-958b-438a-b808-2b9007d74537-cert\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764802 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764818 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzhf5\" (UniqueName: \"kubernetes.io/projected/6d33820a-6276-4e72-820e-27876200d322-kube-api-access-pzhf5\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764832 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-serving-cert\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764847 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-kube-api-access-j6sht\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764864 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7dsr\" (UniqueName: \"kubernetes.io/projected/e75d1cca-da41-4ef9-9fff-3bd99754c344-kube-api-access-d7dsr\") pod \"migrator-59844c95c7-dnqg5\" (UID: \"e75d1cca-da41-4ef9-9fff-3bd99754c344\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764882 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-cabundle\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764934 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvwpl\" (UniqueName: \"kubernetes.io/projected/0868165a-d1ad-4853-8318-53034de4b53b-kube-api-access-zvwpl\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764950 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-config-volume\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764968 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p98kp\" (UniqueName: \"kubernetes.io/projected/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-kube-api-access-p98kp\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.764992 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k4rx\" (UniqueName: \"kubernetes.io/projected/345a87c1-958b-438a-b808-2b9007d74537-kube-api-access-4k4rx\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q442s\" (UniqueName: \"kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765055 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-key\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765071 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765086 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-srv-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765107 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4nq\" (UniqueName: \"kubernetes.io/projected/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-kube-api-access-6m4nq\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765124 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-csi-data-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765144 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-certs\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765158 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765329 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-config\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.765988 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: E1002 10:54:37.766257 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.266245096 +0000 UTC m=+160.320471536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.766784 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.767009 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-registration-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.769021 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-config-volume\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.769261 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-cabundle\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.770022 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-csi-data-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.770221 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-mountpoint-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.771964 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-socket-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.775955 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-plugins-dir\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.777249 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-srv-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.778250 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-serving-cert\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.787387 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0868165a-d1ad-4853-8318-53034de4b53b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.790206 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.791712 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gg8df"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.794415 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.795145 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-certs\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.795575 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-profile-collector-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.796244 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d33820a-6276-4e72-820e-27876200d322-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.802809 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cfd2757a-f362-4ecc-8396-5c8d667f1af7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.805852 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9993f91a-910b-48d3-b6cd-1291630cf77b-signing-key\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.805877 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-node-bootstrap-token\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.805852 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb194ce7-018a-43ac-b414-acd2cc8e60d3-srv-cert\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.805947 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mmjw\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.806195 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-metrics-tls\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.806329 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.806492 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.806963 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/345a87c1-958b-438a-b808-2b9007d74537-cert\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.817026 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bksx\" (UniqueName: \"kubernetes.io/projected/ab8fd399-d4b3-40de-a9cb-fa5481f69c64-kube-api-access-2bksx\") pod \"packageserver-d55dfcdfc-5mlsz\" (UID: \"ab8fd399-d4b3-40de-a9cb-fa5481f69c64\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.818275 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsprr\" (UniqueName: \"kubernetes.io/projected/9b35ed5d-5bc5-4df6-81f4-89e807ae69ed-kube-api-access-hsprr\") pod \"router-default-5444994796-jnthj\" (UID: \"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed\") " pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.819401 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" Oct 02 10:54:37 crc kubenswrapper[4751]: W1002 10:54:37.825338 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2357ddd4_8799_463c_82e6_e98fe53e328b.slice/crio-9d842e26c57ad65cec90d7451d4c774f967a827c1c597272f0270636f7bc8f75 WatchSource:0}: Error finding container 9d842e26c57ad65cec90d7451d4c774f967a827c1c597272f0270636f7bc8f75: Status 404 returned error can't find the container with id 9d842e26c57ad65cec90d7451d4c774f967a827c1c597272f0270636f7bc8f75 Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.835620 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5891fee9-8bb0-490a-8b03-7da135d49156-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lh9jw\" (UID: \"5891fee9-8bb0-490a-8b03-7da135d49156\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: W1002 10:54:37.842827 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod763d7fee_c2ed_4f24_8ff3_27f308054771.slice/crio-a6d0aa3fd72e15866b0227af1a6c62d3dce55bad53dce6d5ccb68322c6cf4e54 WatchSource:0}: Error finding container a6d0aa3fd72e15866b0227af1a6c62d3dce55bad53dce6d5ccb68322c6cf4e54: Status 404 returned error can't find the container with id a6d0aa3fd72e15866b0227af1a6c62d3dce55bad53dce6d5ccb68322c6cf4e54 Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.846762 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mrvdc"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.863098 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97875ab4-12f1-45da-af39-353e06064013-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lt87p\" (UID: \"97875ab4-12f1-45da-af39-353e06064013\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.870042 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:37 crc kubenswrapper[4751]: E1002 10:54:37.870548 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.37053346 +0000 UTC m=+160.424759910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.873756 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2s8q\" (UniqueName: \"kubernetes.io/projected/bc181813-c51d-496e-ba0d-4f03f544d9d2-kube-api-access-l2s8q\") pod \"cluster-image-registry-operator-dc59b4c8b-qjr62\" (UID: \"bc181813-c51d-496e-ba0d-4f03f544d9d2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.887447 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.900770 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2t6m9"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.903650 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68s6f\" (UniqueName: \"kubernetes.io/projected/f14b663f-94ad-44f2-82a8-ec6d53159146-kube-api-access-68s6f\") pod \"downloads-7954f5f757-2dt22\" (UID: \"f14b663f-94ad-44f2-82a8-ec6d53159146\") " pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.914378 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5mv4\" (UniqueName: \"kubernetes.io/projected/d278f4ff-d1f2-4710-a8f4-e87a429062a9-kube-api-access-p5mv4\") pod \"ingress-operator-5b745b69d9-9q8q2\" (UID: \"d278f4ff-d1f2-4710-a8f4-e87a429062a9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.932370 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.933876 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trmgf\" (UniqueName: \"kubernetes.io/projected/2061adbc-fb55-4f79-8d65-93729fb4afb1-kube-api-access-trmgf\") pod \"etcd-operator-b45778765-ks9q4\" (UID: \"2061adbc-fb55-4f79-8d65-93729fb4afb1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.940304 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.948473 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.959418 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79zgg\" (UniqueName: \"kubernetes.io/projected/0acce769-aae9-45be-b613-ff87c0d23453-kube-api-access-79zgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-ctf2s\" (UID: \"0acce769-aae9-45be-b613-ff87c0d23453\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.988142 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.988226 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.988225 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.988371 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" Oct 02 10:54:37 crc kubenswrapper[4751]: E1002 10:54:37.988520 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.488506829 +0000 UTC m=+160.542733279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.996463 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp"] Oct 02 10:54:37 crc kubenswrapper[4751]: I1002 10:54:37.996815 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.011241 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.015061 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.015886 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtvxx\" (UniqueName: \"kubernetes.io/projected/8add3b3a-0164-4e25-9caf-ec05032c9b90-kube-api-access-wtvxx\") pod \"machine-config-controller-84d6567774-mcfjr\" (UID: \"8add3b3a-0164-4e25-9caf-ec05032c9b90\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.031719 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j8ct\" (UniqueName: \"kubernetes.io/projected/eb194ce7-018a-43ac-b414-acd2cc8e60d3-kube-api-access-4j8ct\") pod \"catalog-operator-68c6474976-pxm9p\" (UID: \"eb194ce7-018a-43ac-b414-acd2cc8e60d3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.031970 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.048428 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7lhjm"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.060511 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgszr\" (UniqueName: \"kubernetes.io/projected/9993f91a-910b-48d3-b6cd-1291630cf77b-kube-api-access-dgszr\") pod \"service-ca-9c57cc56f-swmnp\" (UID: \"9993f91a-910b-48d3-b6cd-1291630cf77b\") " pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.070061 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.070377 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.073155 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7cl\" (UniqueName: \"kubernetes.io/projected/cfd2757a-f362-4ecc-8396-5c8d667f1af7-kube-api-access-pd7cl\") pod \"multus-admission-controller-857f4d67dd-rr827\" (UID: \"cfd2757a-f362-4ecc-8396-5c8d667f1af7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.074667 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58b8cdd6_dfa9_4e52_8f93_12663a522318.slice/crio-970b4f33695d4b435947cce89faed2a1d79ab284be2cc23e0bf297ca0a7d1d96 WatchSource:0}: Error finding container 970b4f33695d4b435947cce89faed2a1d79ab284be2cc23e0bf297ca0a7d1d96: Status 404 returned error can't find the container with id 970b4f33695d4b435947cce89faed2a1d79ab284be2cc23e0bf297ca0a7d1d96 Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.075707 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b35ed5d_5bc5_4df6_81f4_89e807ae69ed.slice/crio-18e24973b4fc796d42e2cb02d3343b1d7637cf76a02412540e7659b1ecb275c0 WatchSource:0}: Error finding container 18e24973b4fc796d42e2cb02d3343b1d7637cf76a02412540e7659b1ecb275c0: Status 404 returned error can't find the container with id 18e24973b4fc796d42e2cb02d3343b1d7637cf76a02412540e7659b1ecb275c0 Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.081672 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0413ee62_e2ef_4ca0_a21f_2990f11169c3.slice/crio-3fe41d3cf8ab26496760c1c779294ebb07fdd4c9794786b20c7ac39696c64d1c WatchSource:0}: Error finding container 3fe41d3cf8ab26496760c1c779294ebb07fdd4c9794786b20c7ac39696c64d1c: Status 404 returned error can't find the container with id 3fe41d3cf8ab26496760c1c779294ebb07fdd4c9794786b20c7ac39696c64d1c Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.090994 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.091527 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.591511947 +0000 UTC m=+160.645738397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.093821 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4r69k"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.095883 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q442s\" (UniqueName: \"kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s\") pod \"collect-profiles-29323365-wt5rw\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.105054 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae90eabd_ce14_4792_b171_589323678ab0.slice/crio-e487ff1053204235827eda3092ed349aa6b504f040cc9ac31679f346a36252a1 WatchSource:0}: Error finding container e487ff1053204235827eda3092ed349aa6b504f040cc9ac31679f346a36252a1: Status 404 returned error can't find the container with id e487ff1053204235827eda3092ed349aa6b504f040cc9ac31679f346a36252a1 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.115201 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6sht\" (UniqueName: \"kubernetes.io/projected/430e6c19-8e06-4d2e-bad0-6ede3faaf0ad-kube-api-access-j6sht\") pod \"dns-default-bwnql\" (UID: \"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad\") " pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.122698 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.133760 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.135412 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7dsr\" (UniqueName: \"kubernetes.io/projected/e75d1cca-da41-4ef9-9fff-3bd99754c344-kube-api-access-d7dsr\") pod \"migrator-59844c95c7-dnqg5\" (UID: \"e75d1cca-da41-4ef9-9fff-3bd99754c344\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.141088 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded30f5ae_4c45_4928_9ef6_079587f62490.slice/crio-5748fcd4529f6b186ebe3a3b41bc89383dd4f17860480f2a8440d8c1d2d05cb0 WatchSource:0}: Error finding container 5748fcd4529f6b186ebe3a3b41bc89383dd4f17860480f2a8440d8c1d2d05cb0: Status 404 returned error can't find the container with id 5748fcd4529f6b186ebe3a3b41bc89383dd4f17860480f2a8440d8c1d2d05cb0 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.149161 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.153664 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvwpl\" (UniqueName: \"kubernetes.io/projected/0868165a-d1ad-4853-8318-53034de4b53b-kube-api-access-zvwpl\") pod \"olm-operator-6b444d44fb-c4lgh\" (UID: \"0868165a-d1ad-4853-8318-53034de4b53b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.162213 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc181813_c51d_496e_ba0d_4f03f544d9d2.slice/crio-fd52b87c72ce4164233dce304c8cf8e05d82fdde37bc532dba8155b0246c7c7b WatchSource:0}: Error finding container fd52b87c72ce4164233dce304c8cf8e05d82fdde37bc532dba8155b0246c7c7b: Status 404 returned error can't find the container with id fd52b87c72ce4164233dce304c8cf8e05d82fdde37bc532dba8155b0246c7c7b Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.174586 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4nq\" (UniqueName: \"kubernetes.io/projected/50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48-kube-api-access-6m4nq\") pod \"machine-config-server-v4ns6\" (UID: \"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48\") " pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.181822 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.191945 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w4wn\" (UniqueName: \"kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn\") pod \"marketplace-operator-79b997595-ht64t\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.192297 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.192627 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.692610982 +0000 UTC m=+160.746837432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.195135 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.210940 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.214635 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjr4r\" (UniqueName: \"kubernetes.io/projected/9d73c4ff-f50e-48f9-9feb-3b48e7740b16-kube-api-access-bjr4r\") pod \"service-ca-operator-777779d784-gbb5r\" (UID: \"9d73c4ff-f50e-48f9-9feb-3b48e7740b16\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.234739 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k4rx\" (UniqueName: \"kubernetes.io/projected/345a87c1-958b-438a-b808-2b9007d74537-kube-api-access-4k4rx\") pod \"ingress-canary-7q7g2\" (UID: \"345a87c1-958b-438a-b808-2b9007d74537\") " pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.257137 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzhf5\" (UniqueName: \"kubernetes.io/projected/6d33820a-6276-4e72-820e-27876200d322-kube-api-access-pzhf5\") pod \"package-server-manager-789f6589d5-m2m9f\" (UID: \"6d33820a-6276-4e72-820e-27876200d322\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.281359 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p98kp\" (UniqueName: \"kubernetes.io/projected/ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9-kube-api-access-p98kp\") pod \"csi-hostpathplugin-6swj5\" (UID: \"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9\") " pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.293541 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.293902 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.793888082 +0000 UTC m=+160.848114532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.303659 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.309091 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.318267 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.323930 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.347060 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.359575 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.371449 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.375250 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v4ns6" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.387044 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7q7g2" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.391926 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.395378 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.395715 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.895701846 +0000 UTC m=+160.949928296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.401252 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" event={"ID":"bc181813-c51d-496e-ba0d-4f03f544d9d2","Type":"ContainerStarted","Data":"fd52b87c72ce4164233dce304c8cf8e05d82fdde37bc532dba8155b0246c7c7b"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.412403 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" event={"ID":"58b8cdd6-dfa9-4e52-8f93-12663a522318","Type":"ContainerStarted","Data":"970b4f33695d4b435947cce89faed2a1d79ab284be2cc23e0bf297ca0a7d1d96"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.414122 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" event={"ID":"ed30f5ae-4c45-4928-9ef6-079587f62490","Type":"ContainerStarted","Data":"5748fcd4529f6b186ebe3a3b41bc89383dd4f17860480f2a8440d8c1d2d05cb0"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.415511 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" event={"ID":"bd01da30-84c7-4814-a15a-0ea85c1de44f","Type":"ContainerStarted","Data":"ce97180e22250211c4a7e6c5a015754f01d3b11dcf46230fc18ce455b6608e18"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.417328 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" event={"ID":"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a","Type":"ContainerStarted","Data":"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.417355 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" event={"ID":"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a","Type":"ContainerStarted","Data":"54d9c7286ed283f9cb8050e46d4ef9626d0aa3c80ecaf04be2249f1f62944149"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.417682 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.419793 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.419840 4751 generic.go:334] "Generic (PLEG): container finished" podID="763d7fee-c2ed-4f24-8ff3-27f308054771" containerID="fbb2174cd1eba32e850640189894fdf2c3891b3f9e8ff3fdde4077f679f50e0f" exitCode=0 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.419912 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" event={"ID":"763d7fee-c2ed-4f24-8ff3-27f308054771","Type":"ContainerDied","Data":"fbb2174cd1eba32e850640189894fdf2c3891b3f9e8ff3fdde4077f679f50e0f"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.419969 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" event={"ID":"763d7fee-c2ed-4f24-8ff3-27f308054771","Type":"ContainerStarted","Data":"a6d0aa3fd72e15866b0227af1a6c62d3dce55bad53dce6d5ccb68322c6cf4e54"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.421375 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" event={"ID":"acd9ab77-c505-4404-95df-a0050d7e8a00","Type":"ContainerStarted","Data":"01b4ac69b0def616b8c767b3f70136a9cafd1283637e3cd8f376709e7341b839"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.423890 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.424461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jnthj" event={"ID":"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed","Type":"ContainerStarted","Data":"18e24973b4fc796d42e2cb02d3343b1d7637cf76a02412540e7659b1ecb275c0"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.425998 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k2g7v" event={"ID":"b419cc91-b70e-4108-a04a-761f0ce61313","Type":"ContainerStarted","Data":"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.426105 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k2g7v" event={"ID":"b419cc91-b70e-4108-a04a-761f0ce61313","Type":"ContainerStarted","Data":"6bcf9d119ff9dbfcf8ef5894d22fdc019bf78c86d48db199f5b6dae253825bed"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.426900 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" event={"ID":"2d419bc9-8000-4cdc-bc74-be1147be91e9","Type":"ContainerStarted","Data":"d0127c188b8dd55e2ac09abc19b0dad3ff55089c126e50565c9601461e895d12"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.428518 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" event={"ID":"ae90eabd-ce14-4792-b171-589323678ab0","Type":"ContainerStarted","Data":"e487ff1053204235827eda3092ed349aa6b504f040cc9ac31679f346a36252a1"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.431821 4751 generic.go:334] "Generic (PLEG): container finished" podID="f33d078a-62a0-4f6e-8efb-85003d873289" containerID="850d141ea9fdc2de3a73e209523bc82ae033cf06851f7f1367b7b88ec441dcf8" exitCode=0 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.431934 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" event={"ID":"f33d078a-62a0-4f6e-8efb-85003d873289","Type":"ContainerDied","Data":"850d141ea9fdc2de3a73e209523bc82ae033cf06851f7f1367b7b88ec441dcf8"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.432001 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" event={"ID":"f33d078a-62a0-4f6e-8efb-85003d873289","Type":"ContainerStarted","Data":"53da1087eadc1a361e3ee11138ea70ab88b298b7ed8cfaa7e842a5ed210f3c90"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.441659 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" event={"ID":"176d9f58-b454-4171-8297-f295b8650956","Type":"ContainerStarted","Data":"48676b1ae29c640ed2d39e576d3b3691cf2a89bbb08a4d8b7dec70f1dee73024"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.441711 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" event={"ID":"176d9f58-b454-4171-8297-f295b8650956","Type":"ContainerStarted","Data":"b402fe7c88b9a8932d7f7f3f83354a955f192358e860cb048e114464fbb0f99b"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.443957 4751 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ztmtr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.443992 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.444333 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" event={"ID":"0413ee62-e2ef-4ca0-a21f-2990f11169c3","Type":"ContainerStarted","Data":"3fe41d3cf8ab26496760c1c779294ebb07fdd4c9794786b20c7ac39696c64d1c"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.452009 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" event={"ID":"40a6cb79-0b3a-49b0-9b5c-580030f9978d","Type":"ContainerStarted","Data":"b50e1c9d26d42085f0f9dd261e38bbdf837bee4b842c6d85e9e4cf08aa610840"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.454570 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" event={"ID":"5fcfe397-0564-4c53-b2c2-0b1bc0796269","Type":"ContainerStarted","Data":"8e131db94432d673a5e617e66ff7f80694c837af6f7f9b1e5554e746d8af31d0"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.454605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" event={"ID":"5fcfe397-0564-4c53-b2c2-0b1bc0796269","Type":"ContainerStarted","Data":"964b770ee7a5418317c61b5521619fd463ece1d0867ef7c446277f418baf5f07"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.455761 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" event={"ID":"720be237-ff1f-4ed2-b55b-6e46c976cce8","Type":"ContainerStarted","Data":"0a4cdd08db7ff21f6d3e17fb58a7d3773ffc15da85c14596241d791a30425a87"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.459744 4751 generic.go:334] "Generic (PLEG): container finished" podID="2357ddd4-8799-463c-82e6-e98fe53e328b" containerID="43c2040e99db31245bf01987c114453c889fe66bbd2d1d19281a255ad4c68361" exitCode=0 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.459816 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" event={"ID":"2357ddd4-8799-463c-82e6-e98fe53e328b","Type":"ContainerDied","Data":"43c2040e99db31245bf01987c114453c889fe66bbd2d1d19281a255ad4c68361"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.459862 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" event={"ID":"2357ddd4-8799-463c-82e6-e98fe53e328b","Type":"ContainerStarted","Data":"9d842e26c57ad65cec90d7451d4c774f967a827c1c597272f0270636f7bc8f75"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.461996 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" event={"ID":"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a","Type":"ContainerStarted","Data":"0b871a0d6bacd1b345d465c7b5733d693d3f499d8695e045c2d021609353c0b8"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.462040 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" event={"ID":"2b6741e3-53aa-4f28-a2af-8c6a0a8dae5a","Type":"ContainerStarted","Data":"d11b47b6f33df801611b125afe23295089799a03e2ddde9a330e6253becc9046"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.464264 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" event={"ID":"d5057ef7-7363-4d5c-b72c-d7473c37696d","Type":"ContainerStarted","Data":"19389cc78de5bfbdca76b2bbc9f784c2ad6cc595e14147849936b120cb9675b7"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.464294 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" event={"ID":"d5057ef7-7363-4d5c-b72c-d7473c37696d","Type":"ContainerStarted","Data":"2a157412faf719f879cc97dedb11fbe46d5a6737da016b89d9f0b6b6691b5c86"} Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.464944 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.471614 4751 patch_prober.go:28] interesting pod/console-operator-58897d9998-mrvdc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.471685 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" podUID="d5057ef7-7363-4d5c-b72c-d7473c37696d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.496013 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.496212 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:38.996189594 +0000 UTC m=+161.050416054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.496544 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.500948 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.000928897 +0000 UTC m=+161.055155347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.599230 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.600417 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.100395285 +0000 UTC m=+161.154621735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.605130 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ks9q4"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.607489 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.652588 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p"] Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.678779 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2061adbc_fb55_4f79_8d65_93729fb4afb1.slice/crio-8dd28b5d963b18cffb382175468a33fb38e01e527cffebd2b382335f9883356d WatchSource:0}: Error finding container 8dd28b5d963b18cffb382175468a33fb38e01e527cffebd2b382335f9883356d: Status 404 returned error can't find the container with id 8dd28b5d963b18cffb382175468a33fb38e01e527cffebd2b382335f9883356d Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.678795 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.702800 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.703188 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.203157407 +0000 UTC m=+161.257383857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.732553 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97875ab4_12f1_45da_af39_353e06064013.slice/crio-3fc866931d03b620f79aabfad4b66a08b121204fc746d7221765cb114160ef35 WatchSource:0}: Error finding container 3fc866931d03b620f79aabfad4b66a08b121204fc746d7221765cb114160ef35: Status 404 returned error can't find the container with id 3fc866931d03b620f79aabfad4b66a08b121204fc746d7221765cb114160ef35 Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.753399 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.804353 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.804490 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.304466827 +0000 UTC m=+161.358693277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.804672 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.805053 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.305041643 +0000 UTC m=+161.359268103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.814730 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-swmnp"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.855032 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" podStartSLOduration=128.855009749 podStartE2EDuration="2m8.855009749s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:38.852451977 +0000 UTC m=+160.906678437" watchObservedRunningTime="2025-10-02 10:54:38.855009749 +0000 UTC m=+160.909236209" Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.909062 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:38 crc kubenswrapper[4751]: E1002 10:54:38.909794 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.40976949 +0000 UTC m=+161.463995950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.910139 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9993f91a_910b_48d3_b6cd_1291630cf77b.slice/crio-8a6e3a430d766489fc3a3c5e3c0e4f437e4dab7732e2f867d7d176465ea2d6e4 WatchSource:0}: Error finding container 8a6e3a430d766489fc3a3c5e3c0e4f437e4dab7732e2f867d7d176465ea2d6e4: Status 404 returned error can't find the container with id 8a6e3a430d766489fc3a3c5e3c0e4f437e4dab7732e2f867d7d176465ea2d6e4 Oct 02 10:54:38 crc kubenswrapper[4751]: W1002 10:54:38.916115 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab8fd399_d4b3_40de_a9cb_fa5481f69c64.slice/crio-8ab29a5cf8b8b80cb1084d5d136b452f24a75a96775f9b0eede30d25719375ef WatchSource:0}: Error finding container 8ab29a5cf8b8b80cb1084d5d136b452f24a75a96775f9b0eede30d25719375ef: Status 404 returned error can't find the container with id 8ab29a5cf8b8b80cb1084d5d136b452f24a75a96775f9b0eede30d25719375ef Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.984756 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2dt22"] Oct 02 10:54:38 crc kubenswrapper[4751]: I1002 10:54:38.994300 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.008241 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.034987 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.035418 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.535403315 +0000 UTC m=+161.589629765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.061411 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2t6m9" podStartSLOduration=129.061384256 podStartE2EDuration="2m9.061384256s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:39.055652585 +0000 UTC m=+161.109879055" watchObservedRunningTime="2025-10-02 10:54:39.061384256 +0000 UTC m=+161.115610726" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.075906 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s"] Oct 02 10:54:39 crc kubenswrapper[4751]: W1002 10:54:39.113851 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf14b663f_94ad_44f2_82a8_ec6d53159146.slice/crio-f89797027a70cddf3e72af7c06a2d0287a0b9dca590f30cac0217915616f6250 WatchSource:0}: Error finding container f89797027a70cddf3e72af7c06a2d0287a0b9dca590f30cac0217915616f6250: Status 404 returned error can't find the container with id f89797027a70cddf3e72af7c06a2d0287a0b9dca590f30cac0217915616f6250 Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.136332 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.136732 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.636712985 +0000 UTC m=+161.690939435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.137010 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.137481 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.637471077 +0000 UTC m=+161.691697527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: W1002 10:54:39.267277 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0acce769_aae9_45be_b613_ff87c0d23453.slice/crio-c3eafe3978616e706ca00c8136f70631d0b8e7e213b56f9632ae939cb9e3ed27 WatchSource:0}: Error finding container c3eafe3978616e706ca00c8136f70631d0b8e7e213b56f9632ae939cb9e3ed27: Status 404 returned error can't find the container with id c3eafe3978616e706ca00c8136f70631d0b8e7e213b56f9632ae939cb9e3ed27 Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.268828 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.269203 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.769184513 +0000 UTC m=+161.823410963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.280557 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.287214 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.334441 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bwnql"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.341624 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7q7g2"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.344901 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6swj5"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.347002 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.349013 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.369960 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.370310 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.870298198 +0000 UTC m=+161.924524648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.405892 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.480695 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.481077 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:39.981062264 +0000 UTC m=+162.035288714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.485797 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rr827"] Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.528734 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" event={"ID":"e75d1cca-da41-4ef9-9fff-3bd99754c344","Type":"ContainerStarted","Data":"aed36b98f120a68e8711a73faf3f906be3538dbe94a246bb24acbf50bb92292c"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.542641 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" event={"ID":"0868165a-d1ad-4853-8318-53034de4b53b","Type":"ContainerStarted","Data":"5de18f63dfd0d7cedc323999e4160dda4514898fdef6b709095292036c448086"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.582229 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.582554 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.082542909 +0000 UTC m=+162.136769369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.595070 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-c4z9t" podStartSLOduration=129.595052591 podStartE2EDuration="2m9.595052591s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:39.593013924 +0000 UTC m=+161.647240404" watchObservedRunningTime="2025-10-02 10:54:39.595052591 +0000 UTC m=+161.649279041" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.618275 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" event={"ID":"9993f91a-910b-48d3-b6cd-1291630cf77b","Type":"ContainerStarted","Data":"8a6e3a430d766489fc3a3c5e3c0e4f437e4dab7732e2f867d7d176465ea2d6e4"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.618423 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" event={"ID":"58b8cdd6-dfa9-4e52-8f93-12663a522318","Type":"ContainerStarted","Data":"212ec96eba60f14f8be05a4260eed78499d8165357a9978b8d5f3f5e5685bc37"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.620121 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" event={"ID":"3a8d6102-cb02-4184-9ad5-e1fa0a15b406","Type":"ContainerStarted","Data":"7f26b159b38b46136e253da32471f30eee0a10b0f4c4d514fc19f14ededfc0c1"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.623777 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7q7g2" event={"ID":"345a87c1-958b-438a-b808-2b9007d74537","Type":"ContainerStarted","Data":"a0bcc0b6df539e565569f89c46cf361aefa54e7b55ff7e227e12476eb7caf367"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.627436 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" event={"ID":"ab8fd399-d4b3-40de-a9cb-fa5481f69c64","Type":"ContainerStarted","Data":"8ab29a5cf8b8b80cb1084d5d136b452f24a75a96775f9b0eede30d25719375ef"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.629914 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" event={"ID":"9d73c4ff-f50e-48f9-9feb-3b48e7740b16","Type":"ContainerStarted","Data":"ee4853a0f1f68287217c669f7ada34b2dc674f288ee0da8593f68fb679c5c05d"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.632475 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bwnql" event={"ID":"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad","Type":"ContainerStarted","Data":"839606ea79678db75dd18cf3cea71c0460f97b43a901f5002bbaf930892801d9"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.634897 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jnthj" event={"ID":"9b35ed5d-5bc5-4df6-81f4-89e807ae69ed","Type":"ContainerStarted","Data":"917b46b641432cc175f2dc10d1ad1cad878926eef08d0438dacbfad0985dde4b"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.636781 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" event={"ID":"bc181813-c51d-496e-ba0d-4f03f544d9d2","Type":"ContainerStarted","Data":"17af16173ae9443369177f7697222961fee81ca4c1dc7c14800b26a238cbcf04"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.639115 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" event={"ID":"2061adbc-fb55-4f79-8d65-93729fb4afb1","Type":"ContainerStarted","Data":"8dd28b5d963b18cffb382175468a33fb38e01e527cffebd2b382335f9883356d"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.645801 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" event={"ID":"eb194ce7-018a-43ac-b414-acd2cc8e60d3","Type":"ContainerStarted","Data":"d9b9a21ec25c0eac18a68cf3a41b76b16cb0e204e8fe5e59573699b42046e234"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.645832 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" event={"ID":"eb194ce7-018a-43ac-b414-acd2cc8e60d3","Type":"ContainerStarted","Data":"19172958c8fd6b81719b342af7261f1514915ebbea6af5f99897f888d89cc593"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.646574 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.654504 4751 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-pxm9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.654580 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" podUID="eb194ce7-018a-43ac-b414-acd2cc8e60d3" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.667084 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" event={"ID":"ed30f5ae-4c45-4928-9ef6-079587f62490","Type":"ContainerStarted","Data":"6c3b32dec0f6219ab7926ebab0f84e61566365e3432ba128b532896fecc7dbb2"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.675627 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" event={"ID":"ae90eabd-ce14-4792-b171-589323678ab0","Type":"ContainerStarted","Data":"6fffcafc813677e7054e3d83b86abcd38d3626af40c5a16d0dcaa1605feb0fc4"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.683395 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" event={"ID":"d278f4ff-d1f2-4710-a8f4-e87a429062a9","Type":"ContainerStarted","Data":"650146d503bb8f6829b326e5499c1cfc2e57768b8eae83fad963f7e9310c4e84"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.683428 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" event={"ID":"d278f4ff-d1f2-4710-a8f4-e87a429062a9","Type":"ContainerStarted","Data":"ab8d610876ecddc02e1537f8c5d79404f92899d84ebc3c0c6bc52485560ca644"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.688418 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.689202 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.189131978 +0000 UTC m=+162.243358428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.690015 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" podStartSLOduration=129.690001923 podStartE2EDuration="2m9.690001923s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:39.689390926 +0000 UTC m=+161.743617366" watchObservedRunningTime="2025-10-02 10:54:39.690001923 +0000 UTC m=+161.744228373" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.690224 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-k2g7v" podStartSLOduration=128.690220309 podStartE2EDuration="2m8.690220309s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:39.659128044 +0000 UTC m=+161.713354504" watchObservedRunningTime="2025-10-02 10:54:39.690220309 +0000 UTC m=+161.744446759" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.696334 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" event={"ID":"f33d078a-62a0-4f6e-8efb-85003d873289","Type":"ContainerStarted","Data":"8fe36dba87142a9511dfb76b452ea6c40a7225602b8560756bb1f8b15b00bb97"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.696593 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.699047 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" event={"ID":"0acce769-aae9-45be-b613-ff87c0d23453","Type":"ContainerStarted","Data":"c3eafe3978616e706ca00c8136f70631d0b8e7e213b56f9632ae939cb9e3ed27"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.704562 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" event={"ID":"6d33820a-6276-4e72-820e-27876200d322","Type":"ContainerStarted","Data":"4601ebeea5b8d5c2b2669142affa85c27003a40c7f4e959b54cd62afe7ce74fa"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.708558 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2dt22" event={"ID":"f14b663f-94ad-44f2-82a8-ec6d53159146","Type":"ContainerStarted","Data":"f89797027a70cddf3e72af7c06a2d0287a0b9dca590f30cac0217915616f6250"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.710543 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" event={"ID":"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9","Type":"ContainerStarted","Data":"96aca3ccc6aff91db11c25b5092213f82cd3b9486bcef62f6381b0bb97b78ea7"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.712705 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v4ns6" event={"ID":"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48","Type":"ContainerStarted","Data":"e1aec5bcfa133f764ab178acaf181597f8a0d469a73480d89c0f3497660a0b3d"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.712738 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v4ns6" event={"ID":"50ef6015-ffb2-4cd3-87e7-6a2fcfcb4e48","Type":"ContainerStarted","Data":"c6b7edf88dde9898e3df5db5d19c325a9a759893d07001a8c288209a7d5eff1d"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.730697 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" event={"ID":"720be237-ff1f-4ed2-b55b-6e46c976cce8","Type":"ContainerStarted","Data":"9388c8a723340c86978b732f15aeb515a42f192b33d516d12bb50a8f2bb578e9"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.737651 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" event={"ID":"97875ab4-12f1-45da-af39-353e06064013","Type":"ContainerStarted","Data":"3fc866931d03b620f79aabfad4b66a08b121204fc746d7221765cb114160ef35"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.740412 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" event={"ID":"bd01da30-84c7-4814-a15a-0ea85c1de44f","Type":"ContainerStarted","Data":"30b7d0fe93294a88efb424d5ec55c92418939d9fcfbe23e31c6712b2ca7ff4db"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.744367 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" event={"ID":"2d419bc9-8000-4cdc-bc74-be1147be91e9","Type":"ContainerStarted","Data":"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.746633 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.766042 4751 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fppx5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.766097 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.766159 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" event={"ID":"5891fee9-8bb0-490a-8b03-7da135d49156","Type":"ContainerStarted","Data":"7faaed9d87f1f965394c219ed10c9586d990e1f2b385bc04c1eddeb2f171e4f4"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.781238 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" event={"ID":"8add3b3a-0164-4e25-9caf-ec05032c9b90","Type":"ContainerStarted","Data":"1b169a723309eba9276dc68e02e313f66a67d191f5f5145f61ad19c6fe0db1c1"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.787607 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" event={"ID":"0413ee62-e2ef-4ca0-a21f-2990f11169c3","Type":"ContainerStarted","Data":"d106657bb3f4f49fbaee52e78aad5ae734b5d943d5215ecf4f9668e6294203d3"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.791301 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" event={"ID":"acd9ab77-c505-4404-95df-a0050d7e8a00","Type":"ContainerStarted","Data":"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.792087 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.793060 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.793210 4751 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-flnxn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.793276 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.793771 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" event={"ID":"40a6cb79-0b3a-49b0-9b5c-580030f9978d","Type":"ContainerStarted","Data":"1a3472849dd933b518ea96b5503d17756d9dd767409813992cb64095842ef624"} Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.794984 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.294969936 +0000 UTC m=+162.349196386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.796793 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" event={"ID":"5fcfe397-0564-4c53-b2c2-0b1bc0796269","Type":"ContainerStarted","Data":"e01f9b50fde7aa1b5c9e5d09acda1058caa323722a5905ae90caa0f373749779"} Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.799388 4751 patch_prober.go:28] interesting pod/console-operator-58897d9998-mrvdc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.799426 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" podUID="d5057ef7-7363-4d5c-b72c-d7473c37696d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.807159 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.894854 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.895195 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.395158515 +0000 UTC m=+162.449384965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.898102 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:39 crc kubenswrapper[4751]: E1002 10:54:39.898789 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.398777987 +0000 UTC m=+162.453004427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.940874 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.943814 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 02 10:54:39 crc kubenswrapper[4751]: I1002 10:54:39.943874 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:39.999968 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.000810 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.500790357 +0000 UTC m=+162.555016807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.095273 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" podStartSLOduration=130.095235515 podStartE2EDuration="2m10.095235515s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.093870746 +0000 UTC m=+162.148097196" watchObservedRunningTime="2025-10-02 10:54:40.095235515 +0000 UTC m=+162.149461975" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.096812 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-p5jlr" podStartSLOduration=129.096800619 podStartE2EDuration="2m9.096800619s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.05883654 +0000 UTC m=+162.113063000" watchObservedRunningTime="2025-10-02 10:54:40.096800619 +0000 UTC m=+162.151027079" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.101512 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.101771 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.601760608 +0000 UTC m=+162.655987058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.187158 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" podStartSLOduration=129.187141921 podStartE2EDuration="2m9.187141921s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.186618356 +0000 UTC m=+162.240844806" watchObservedRunningTime="2025-10-02 10:54:40.187141921 +0000 UTC m=+162.241368371" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.205605 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.205893 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.705867517 +0000 UTC m=+162.760093967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.206594 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.207037 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.70701846 +0000 UTC m=+162.761244900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.234937 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" podStartSLOduration=129.234920345 podStartE2EDuration="2m9.234920345s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.234080701 +0000 UTC m=+162.288307151" watchObservedRunningTime="2025-10-02 10:54:40.234920345 +0000 UTC m=+162.289146795" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.311549 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.311666 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.811642774 +0000 UTC m=+162.865869224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.312081 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.312476 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.812466167 +0000 UTC m=+162.866692617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.412927 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.413186 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.913130729 +0000 UTC m=+162.967357169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.413529 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.413964 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:40.913953992 +0000 UTC m=+162.968180442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.520606 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.520843 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.020817349 +0000 UTC m=+163.075043799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.521574 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.522038 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.022027353 +0000 UTC m=+163.076253803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.554373 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qjr62" podStartSLOduration=129.554353702 podStartE2EDuration="2m9.554353702s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.552110169 +0000 UTC m=+162.606336619" watchObservedRunningTime="2025-10-02 10:54:40.554353702 +0000 UTC m=+162.608580152" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.626467 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.626779 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.12676522 +0000 UTC m=+163.180991670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.644882 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rqr28" podStartSLOduration=129.644855079 podStartE2EDuration="2m9.644855079s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.596118148 +0000 UTC m=+162.650344598" watchObservedRunningTime="2025-10-02 10:54:40.644855079 +0000 UTC m=+162.699081529" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.720728 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zzmsd" podStartSLOduration=129.720696063 podStartE2EDuration="2m9.720696063s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.650512918 +0000 UTC m=+162.704739378" watchObservedRunningTime="2025-10-02 10:54:40.720696063 +0000 UTC m=+162.774922513" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.727755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.728105 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.228087731 +0000 UTC m=+163.282314181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.763512 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jnthj" podStartSLOduration=129.763493706 podStartE2EDuration="2m9.763493706s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.720577869 +0000 UTC m=+162.774804329" watchObservedRunningTime="2025-10-02 10:54:40.763493706 +0000 UTC m=+162.817720156" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.786887 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" podStartSLOduration=129.786869924 podStartE2EDuration="2m9.786869924s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.763616489 +0000 UTC m=+162.817842939" watchObservedRunningTime="2025-10-02 10:54:40.786869924 +0000 UTC m=+162.841096374" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.815100 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7t5dg" podStartSLOduration=130.815081047 podStartE2EDuration="2m10.815081047s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.811337742 +0000 UTC m=+162.865564192" watchObservedRunningTime="2025-10-02 10:54:40.815081047 +0000 UTC m=+162.869307497" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.817652 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-v4ns6" podStartSLOduration=5.817643819 podStartE2EDuration="5.817643819s" podCreationTimestamp="2025-10-02 10:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.788643863 +0000 UTC m=+162.842870323" watchObservedRunningTime="2025-10-02 10:54:40.817643819 +0000 UTC m=+162.871870259" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.828508 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.829099 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.32904368 +0000 UTC m=+163.383270130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.883416 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" event={"ID":"8add3b3a-0164-4e25-9caf-ec05032c9b90","Type":"ContainerStarted","Data":"c82f318dee6a4cd77b8c2c206fe506171828ed927559897da542ae7329e6f846"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.913256 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" event={"ID":"763d7fee-c2ed-4f24-8ff3-27f308054771","Type":"ContainerStarted","Data":"180143cdb98df496caea3ffaa0604e3e8a6b2c0d44d1b9f1f55f5d2bdf9ec85b"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.930232 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:40 crc kubenswrapper[4751]: E1002 10:54:40.930539 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.430527376 +0000 UTC m=+163.484753826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.940072 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bwnql" event={"ID":"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad","Type":"ContainerStarted","Data":"47409b72dab3b39db5b096f65a25aa0b2d816c928d9044745e6e3550a94f7642"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.949185 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:40 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:40 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:40 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.949259 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.959969 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" event={"ID":"d278f4ff-d1f2-4710-a8f4-e87a429062a9","Type":"ContainerStarted","Data":"99befdf4d874890488dba6837e15a6b438bc8fe847e5b277a1864eff57d3d365"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.972806 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" event={"ID":"ab8fd399-d4b3-40de-a9cb-fa5481f69c64","Type":"ContainerStarted","Data":"fbc070c7ce19607add01edbea2830999d64f01ce92d8d02ff67609bef08cf0a5"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.973257 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.974425 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" event={"ID":"5891fee9-8bb0-490a-8b03-7da135d49156","Type":"ContainerStarted","Data":"52e077fed2a38bc64ca86d3a14c59efcc107f88011bb2b7ccb76ccd9a7cd0ae6"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.980693 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" event={"ID":"9993f91a-910b-48d3-b6cd-1291630cf77b","Type":"ContainerStarted","Data":"ce1ab0e2badb14e30fb370a12bcc579d2652b8ed2aff8ac70889db6fba52f5cc"} Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.991532 4751 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5mlsz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.991612 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" podUID="ab8fd399-d4b3-40de-a9cb-fa5481f69c64" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.992193 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9q8q2" podStartSLOduration=129.99215593 podStartE2EDuration="2m9.99215593s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:40.991977825 +0000 UTC m=+163.046204285" watchObservedRunningTime="2025-10-02 10:54:40.99215593 +0000 UTC m=+163.046382380" Oct 02 10:54:40 crc kubenswrapper[4751]: I1002 10:54:40.993495 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" event={"ID":"97875ab4-12f1-45da-af39-353e06064013","Type":"ContainerStarted","Data":"5c4d02f1f6b2b9744a65ce62a16cc02fba810da7cc78ed62b16c6e31cb9187b1"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.007022 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2dt22" event={"ID":"f14b663f-94ad-44f2-82a8-ec6d53159146","Type":"ContainerStarted","Data":"8851fec37cab6527ab80b195d0586c1074d694be91a85e79d8d621d8f1412318"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.007598 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.011850 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.011900 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.014914 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" event={"ID":"2061adbc-fb55-4f79-8d65-93729fb4afb1","Type":"ContainerStarted","Data":"4cf2d3b83137c67a22ab15142e53c82d463638333581663ab718bed8225e8021"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.028631 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" event={"ID":"58b8cdd6-dfa9-4e52-8f93-12663a522318","Type":"ContainerStarted","Data":"f56eafe269062124f96bb78fc59f0213b24f2cd01aa9b4ae31373feccc7ac48c"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.030088 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lh9jw" podStartSLOduration=130.030072546 podStartE2EDuration="2m10.030072546s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.028506602 +0000 UTC m=+163.082733052" watchObservedRunningTime="2025-10-02 10:54:41.030072546 +0000 UTC m=+163.084298996" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.034335 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.034671 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.534635965 +0000 UTC m=+163.588862415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.038598 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" event={"ID":"3a8d6102-cb02-4184-9ad5-e1fa0a15b406","Type":"ContainerStarted","Data":"a5f1a2e8ff506be9ebabee4247602b35369362cb75006f63c2476fc3d6f99f78"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.042759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7q7g2" event={"ID":"345a87c1-958b-438a-b808-2b9007d74537","Type":"ContainerStarted","Data":"e82331b05d2462fd59ea27fff6381d46690e12d4097bc157ad78a1487a513f74"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.045973 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" event={"ID":"cfd2757a-f362-4ecc-8396-5c8d667f1af7","Type":"ContainerStarted","Data":"e9de7b3e928ed18ec0956c8156922cf00d2cd326df4185f8df685e5a763cd728"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.046019 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" event={"ID":"cfd2757a-f362-4ecc-8396-5c8d667f1af7","Type":"ContainerStarted","Data":"a9d27072bbfc013f811aa94b25690d232bed23546e2868f3c82b2992bcecb480"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.054214 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" podStartSLOduration=130.054198855 podStartE2EDuration="2m10.054198855s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.052507498 +0000 UTC m=+163.106733958" watchObservedRunningTime="2025-10-02 10:54:41.054198855 +0000 UTC m=+163.108425305" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.072897 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" event={"ID":"bd01da30-84c7-4814-a15a-0ea85c1de44f","Type":"ContainerStarted","Data":"961c727ddd3237b10551afa6e9435dea81e12924add0b09e14bc3595590efd55"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.076926 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" event={"ID":"e75d1cca-da41-4ef9-9fff-3bd99754c344","Type":"ContainerStarted","Data":"8f67aa098d2091326500dbd0b175b666cc7e26c6aa2f35739c8310efe4e56df9"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.076975 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" event={"ID":"e75d1cca-da41-4ef9-9fff-3bd99754c344","Type":"ContainerStarted","Data":"d069719bb9f0de9bf02e3e7861ba7654bfcf91864df35ef015378ba3803921b8"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.082979 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" event={"ID":"ed30f5ae-4c45-4928-9ef6-079587f62490","Type":"ContainerStarted","Data":"96546e1b5bbb26590a107f5efaa660c0230d92a5841f552a0a8f81b343f91140"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.086612 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" event={"ID":"9d73c4ff-f50e-48f9-9feb-3b48e7740b16","Type":"ContainerStarted","Data":"f45aafffe4ac3b1c7e19748fe4d16a1f375e7ad0be97e9b46ef7d40115c1b366"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.090272 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" event={"ID":"0acce769-aae9-45be-b613-ff87c0d23453","Type":"ContainerStarted","Data":"7c2d623b10d9e3dbfc8fd2d3564b5082bdf086e4951d018a8fa337f6bfbc5ce3"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.099018 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-swmnp" podStartSLOduration=130.098998056 podStartE2EDuration="2m10.098998056s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.071995096 +0000 UTC m=+163.126221546" watchObservedRunningTime="2025-10-02 10:54:41.098998056 +0000 UTC m=+163.153224506" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.103697 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" podStartSLOduration=131.103679527 podStartE2EDuration="2m11.103679527s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.099000266 +0000 UTC m=+163.153226716" watchObservedRunningTime="2025-10-02 10:54:41.103679527 +0000 UTC m=+163.157905977" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.111504 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" event={"ID":"ae90eabd-ce14-4792-b171-589323678ab0","Type":"ContainerStarted","Data":"a255dea4979eaadb430353458d40bf4eb5996b45fd1eee23deda31e712b74343"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.117395 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" event={"ID":"0868165a-d1ad-4853-8318-53034de4b53b","Type":"ContainerStarted","Data":"450af96a599809da28f5fe8dce2983e80b763dee54d6e229db5514cc1a692011"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.118480 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.127978 4751 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c4lgh container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.128601 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" podUID="0868165a-d1ad-4853-8318-53034de4b53b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.134998 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" event={"ID":"6d33820a-6276-4e72-820e-27876200d322","Type":"ContainerStarted","Data":"b49677629b858864bfcd9a9888e67dcf6f3b1be17889a9db01d055bf1761ae64"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.135057 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" event={"ID":"6d33820a-6276-4e72-820e-27876200d322","Type":"ContainerStarted","Data":"1352400b23cd148a5e05838f4338faa65bc13848ff7656db6c7cca8ad20a4bd7"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.135468 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.136394 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.139113 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.639097464 +0000 UTC m=+163.693323914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.160482 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" event={"ID":"4718f397-506a-4690-8942-caecf3a3e6c6","Type":"ContainerStarted","Data":"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.160521 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.160530 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" event={"ID":"4718f397-506a-4690-8942-caecf3a3e6c6","Type":"ContainerStarted","Data":"e152a3e2390d495eb0b65a270c27a973455296e764447a780aa22761d64e65f3"} Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.161601 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2dt22" podStartSLOduration=130.161583947 podStartE2EDuration="2m10.161583947s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.127645952 +0000 UTC m=+163.181872402" watchObservedRunningTime="2025-10-02 10:54:41.161583947 +0000 UTC m=+163.215810397" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.162669 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-k2xtp" podStartSLOduration=130.162663757 podStartE2EDuration="2m10.162663757s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.154422345 +0000 UTC m=+163.208648785" watchObservedRunningTime="2025-10-02 10:54:41.162663757 +0000 UTC m=+163.216890207" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.164538 4751 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ht64t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.164582 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.171725 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.172004 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mrvdc" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.178962 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.179213 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pxm9p" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.198974 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lt87p" podStartSLOduration=130.198958328 podStartE2EDuration="2m10.198958328s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.196037966 +0000 UTC m=+163.250264416" watchObservedRunningTime="2025-10-02 10:54:41.198958328 +0000 UTC m=+163.253184778" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.219028 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7q7g2" podStartSLOduration=6.219013132 podStartE2EDuration="6.219013132s" podCreationTimestamp="2025-10-02 10:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.216562804 +0000 UTC m=+163.270789254" watchObservedRunningTime="2025-10-02 10:54:41.219013132 +0000 UTC m=+163.273239572" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.239635 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ks9q4" podStartSLOduration=130.239613302 podStartE2EDuration="2m10.239613302s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.23491489 +0000 UTC m=+163.289141350" watchObservedRunningTime="2025-10-02 10:54:41.239613302 +0000 UTC m=+163.293839762" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.241180 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.242318 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.742298348 +0000 UTC m=+163.796524838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.304968 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gbb5r" podStartSLOduration=130.30495204 podStartE2EDuration="2m10.30495204s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.304532939 +0000 UTC m=+163.358759389" watchObservedRunningTime="2025-10-02 10:54:41.30495204 +0000 UTC m=+163.359178490" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.345671 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.346239 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.846227052 +0000 UTC m=+163.900453502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.358012 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ctf2s" podStartSLOduration=130.357992213 podStartE2EDuration="2m10.357992213s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.356610404 +0000 UTC m=+163.410836854" watchObservedRunningTime="2025-10-02 10:54:41.357992213 +0000 UTC m=+163.412218663" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.412814 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" podStartSLOduration=130.412795045 podStartE2EDuration="2m10.412795045s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.407588168 +0000 UTC m=+163.461814618" watchObservedRunningTime="2025-10-02 10:54:41.412795045 +0000 UTC m=+163.467021495" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.446287 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.446427 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.94640791 +0000 UTC m=+164.000634360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.446577 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.446853 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:41.946845753 +0000 UTC m=+164.001072203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.514250 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" podStartSLOduration=130.514233469 podStartE2EDuration="2m10.514233469s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.449356213 +0000 UTC m=+163.503582673" watchObservedRunningTime="2025-10-02 10:54:41.514233469 +0000 UTC m=+163.568459919" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.547861 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.548063 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.04803598 +0000 UTC m=+164.102262440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.548142 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.548461 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.048445531 +0000 UTC m=+164.102672031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.572589 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4r69k" podStartSLOduration=130.5725718 podStartE2EDuration="2m10.5725718s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.569752571 +0000 UTC m=+163.623979041" watchObservedRunningTime="2025-10-02 10:54:41.5725718 +0000 UTC m=+163.626798250" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.573703 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7pnb7" podStartSLOduration=131.573697692 podStartE2EDuration="2m11.573697692s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.543301087 +0000 UTC m=+163.597527537" watchObservedRunningTime="2025-10-02 10:54:41.573697692 +0000 UTC m=+163.627924142" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.638653 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" podStartSLOduration=130.638637899 podStartE2EDuration="2m10.638637899s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.635419879 +0000 UTC m=+163.689646329" watchObservedRunningTime="2025-10-02 10:54:41.638637899 +0000 UTC m=+163.692864359" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.649658 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.649830 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.149788383 +0000 UTC m=+164.204014833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.649936 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.650246 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.150236435 +0000 UTC m=+164.204462885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.682483 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7lhjm" podStartSLOduration=130.682465352 podStartE2EDuration="2m10.682465352s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.679660833 +0000 UTC m=+163.733887283" watchObservedRunningTime="2025-10-02 10:54:41.682465352 +0000 UTC m=+163.736691792" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.750866 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.751265 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.251248318 +0000 UTC m=+164.305474768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.852769 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.853072 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.353060522 +0000 UTC m=+164.407286972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.946162 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:41 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:41 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:41 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.946244 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.953660 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.953837 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.453813357 +0000 UTC m=+164.508039807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:41 crc kubenswrapper[4751]: I1002 10:54:41.953928 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:41 crc kubenswrapper[4751]: E1002 10:54:41.954245 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.454232999 +0000 UTC m=+164.508459449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.055300 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.055568 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.555543029 +0000 UTC m=+164.609769479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.156770 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.157211 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.657195499 +0000 UTC m=+164.711421949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.167383 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" event={"ID":"763d7fee-c2ed-4f24-8ff3-27f308054771","Type":"ContainerStarted","Data":"bfbd5e545b1b88b4cd0adcb491c4d1f6d94a72f573dfab70169ead5e08c19393"} Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.169926 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" event={"ID":"2357ddd4-8799-463c-82e6-e98fe53e328b","Type":"ContainerStarted","Data":"bc9548e527ed346f6ad665a233e4b5c0245c08081513ecdd2b1635e38508dfd8"} Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.172709 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bwnql" event={"ID":"430e6c19-8e06-4d2e-bad0-6ede3faaf0ad","Type":"ContainerStarted","Data":"a26d9c48510555f908715cadb851741149c7aa1b37acf44626f5fcaa66da698f"} Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.173121 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.174610 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" event={"ID":"cfd2757a-f362-4ecc-8396-5c8d667f1af7","Type":"ContainerStarted","Data":"853fb0659223649d928635d69d259b1c01ecb3cad4da7d5d5f37a7f52bb63cd7"} Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.177726 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" event={"ID":"8add3b3a-0164-4e25-9caf-ec05032c9b90","Type":"ContainerStarted","Data":"8e88fd47c32c04c8ee5dfb08919cf22654282c2fa3912029771d81eb5c7a581b"} Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183347 4751 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ht64t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183351 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183412 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183505 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183904 4751 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-c4lgh container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.183950 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" podUID="0868165a-d1ad-4853-8318-53034de4b53b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.209908 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" podStartSLOduration=132.209887982 podStartE2EDuration="2m12.209887982s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:42.20626362 +0000 UTC m=+164.260490080" watchObservedRunningTime="2025-10-02 10:54:42.209887982 +0000 UTC m=+164.264114432" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.211366 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dnqg5" podStartSLOduration=131.211360063 podStartE2EDuration="2m11.211360063s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:41.715260265 +0000 UTC m=+163.769486725" watchObservedRunningTime="2025-10-02 10:54:42.211360063 +0000 UTC m=+164.265586533" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.233415 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" podStartSLOduration=131.233402534 podStartE2EDuration="2m11.233402534s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:42.232660153 +0000 UTC m=+164.286886603" watchObservedRunningTime="2025-10-02 10:54:42.233402534 +0000 UTC m=+164.287628984" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.258225 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.258452 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.758425488 +0000 UTC m=+164.812651938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.284004 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rr827" podStartSLOduration=131.283988807 podStartE2EDuration="2m11.283988807s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:42.281752464 +0000 UTC m=+164.335978904" watchObservedRunningTime="2025-10-02 10:54:42.283988807 +0000 UTC m=+164.338215267" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.360898 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.364077 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.86405085 +0000 UTC m=+164.918277300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.418979 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mcfjr" podStartSLOduration=131.418961874 podStartE2EDuration="2m11.418961874s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:42.388382984 +0000 UTC m=+164.442609454" watchObservedRunningTime="2025-10-02 10:54:42.418961874 +0000 UTC m=+164.473188324" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.428052 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.428452 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.433311 4751 patch_prober.go:28] interesting pod/apiserver-76f77b778f-gg8df container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.24:8443/livez\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.433383 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" podUID="763d7fee-c2ed-4f24-8ff3-27f308054771" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.24:8443/livez\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.457533 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.457655 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.464404 4751 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-6qpgb container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.464443 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" podUID="2357ddd4-8799-463c-82e6-e98fe53e328b" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.464929 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.465331 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:42.965318649 +0000 UTC m=+165.019545099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.566207 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.566554 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.066541257 +0000 UTC m=+165.120767707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.667413 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.667697 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.167683013 +0000 UTC m=+165.221909463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.768517 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.768874 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.268858659 +0000 UTC m=+165.323085109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.771338 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bwnql" podStartSLOduration=7.771327479 podStartE2EDuration="7.771327479s" podCreationTimestamp="2025-10-02 10:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:42.484490828 +0000 UTC m=+164.538717288" watchObservedRunningTime="2025-10-02 10:54:42.771327479 +0000 UTC m=+164.825553929" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.771527 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.772397 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.778686 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.786037 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.869851 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.870006 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.369981954 +0000 UTC m=+165.424208404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.870247 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.870315 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.870381 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.870430 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9v7v\" (UniqueName: \"kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.870880 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.370859679 +0000 UTC m=+165.425086169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.944132 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:42 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:42 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:42 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.944204 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.961917 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.962750 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.964452 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.971672 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.971871 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.971908 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.971932 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9v7v\" (UniqueName: \"kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: E1002 10:54:42.972338 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.472289623 +0000 UTC m=+165.526516073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.972478 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.972525 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:42 crc kubenswrapper[4751]: I1002 10:54:42.975311 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.009771 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9v7v\" (UniqueName: \"kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v\") pod \"community-operators-nl7xm\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.072925 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.072998 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.073044 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm26h\" (UniqueName: \"kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.073067 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.073577 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.573556492 +0000 UTC m=+165.627782952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.086417 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.174005 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.174437 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.174518 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm26h\" (UniqueName: \"kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.174549 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.175083 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.675051768 +0000 UTC m=+165.729278218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.175793 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.176014 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.177138 4751 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5mlsz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.177201 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" podUID="ab8fd399-d4b3-40de-a9cb-fa5481f69c64" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.180814 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.181734 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.217561 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.238379 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm26h\" (UniqueName: \"kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h\") pod \"certified-operators-wjgj7\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.242965 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" event={"ID":"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9","Type":"ContainerStarted","Data":"f670f9f063c311cc1b34e73a1aa63dc026d32a09221342dda980be47cf055c89"} Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.244018 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.244053 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.268756 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c4lgh" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.276585 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.280843 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nfl9\" (UniqueName: \"kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.280898 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.280935 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.280977 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.281998 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.781981757 +0000 UTC m=+165.836208197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.383199 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.383816 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.384028 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.384151 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nfl9\" (UniqueName: \"kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.384233 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.384274 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.384973 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.884959764 +0000 UTC m=+165.939186214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.387535 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.389524 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.395853 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.444283 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nfl9\" (UniqueName: \"kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9\") pod \"community-operators-t46qb\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.445253 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6tkns" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.487625 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.488579 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:43.988565569 +0000 UTC m=+166.042792019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.497145 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.497309 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgfdj\" (UniqueName: \"kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.497430 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.502617 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.598550 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.598743 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgfdj\" (UniqueName: \"kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.598796 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.598854 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.600022 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.600034 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.100011895 +0000 UTC m=+166.154238345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.600252 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.640139 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgfdj\" (UniqueName: \"kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj\") pod \"certified-operators-f26gx\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.702072 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.702466 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.202452817 +0000 UTC m=+166.256679267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.760966 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.809261 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.809830 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.309801508 +0000 UTC m=+166.364027958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.841662 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.910884 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:43 crc kubenswrapper[4751]: E1002 10:54:43.911239 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.411227861 +0000 UTC m=+166.465454311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.956211 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:43 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:43 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:43 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:43 crc kubenswrapper[4751]: I1002 10:54:43.956471 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.012563 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.013017 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.512999655 +0000 UTC m=+166.567226105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.020815 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.080788 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:54:44 crc kubenswrapper[4751]: W1002 10:54:44.112809 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0da6eb8_21b5_4e3a_8bdb_2c86c70c4bbc.slice/crio-3cc62933d6c3dfa72573cdc0f1cd54f177f4175a712bc339a5075d9fda783c65 WatchSource:0}: Error finding container 3cc62933d6c3dfa72573cdc0f1cd54f177f4175a712bc339a5075d9fda783c65: Status 404 returned error can't find the container with id 3cc62933d6c3dfa72573cdc0f1cd54f177f4175a712bc339a5075d9fda783c65 Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.114022 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.114353 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.614342446 +0000 UTC m=+166.668568896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.214702 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.214986 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.714971668 +0000 UTC m=+166.769198118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.284215 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerStarted","Data":"e9a808cb4fa72310d707f9601f1c1a4b92a9d6661000a98d3919272a938e06aa"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.290214 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" event={"ID":"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9","Type":"ContainerStarted","Data":"84081bdff1e1e2876b78f0cc4d15cdbb183e618169bcf5642964f6bf4810383f"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.291690 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerStarted","Data":"3cc62933d6c3dfa72573cdc0f1cd54f177f4175a712bc339a5075d9fda783c65"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.296748 4751 generic.go:334] "Generic (PLEG): container finished" podID="3a8d6102-cb02-4184-9ad5-e1fa0a15b406" containerID="a5f1a2e8ff506be9ebabee4247602b35369362cb75006f63c2476fc3d6f99f78" exitCode=0 Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.296866 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" event={"ID":"3a8d6102-cb02-4184-9ad5-e1fa0a15b406","Type":"ContainerDied","Data":"a5f1a2e8ff506be9ebabee4247602b35369362cb75006f63c2476fc3d6f99f78"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.300441 4751 generic.go:334] "Generic (PLEG): container finished" podID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerID="2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9" exitCode=0 Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.301445 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerDied","Data":"2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.301467 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerStarted","Data":"8a9eab9681b4fe9f876425da27391da1bde1d42b2f998a8814fff987e4cc160c"} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.302743 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.316753 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.317054 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.81704318 +0000 UTC m=+166.871269630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.354228 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.421656 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.422396 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:44.922367602 +0000 UTC m=+166.976594052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.523597 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.523886 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.023869618 +0000 UTC m=+167.078096068 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.624020 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.624142 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.124116538 +0000 UTC m=+167.178342988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.624320 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.624583 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.124573791 +0000 UTC m=+167.178800241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.725049 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.725249 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.225221813 +0000 UTC m=+167.279448253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.725401 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.725730 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.225718137 +0000 UTC m=+167.279944587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.750626 4751 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.765027 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.765964 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.768327 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.780431 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.826660 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.826856 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.326828572 +0000 UTC m=+167.381055012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.827259 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.827554 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.327541662 +0000 UTC m=+167.381768112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.928538 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.928717 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gk47\" (UniqueName: \"kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.928749 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.428717669 +0000 UTC m=+167.482944119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.928799 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.929038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.929087 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:44 crc kubenswrapper[4751]: E1002 10:54:44.929450 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 10:54:45.429439209 +0000 UTC m=+167.483665659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qg59s" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.946247 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:44 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:44 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:44 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.946318 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.977884 4751 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T10:54:44.750659009Z","Handler":null,"Name":""} Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.980220 4751 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 10:54:44 crc kubenswrapper[4751]: I1002 10:54:44.980266 4751 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.030699 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.030859 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gk47\" (UniqueName: \"kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.030889 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.030970 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.031424 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.031691 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.034116 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.048877 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gk47\" (UniqueName: \"kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47\") pod \"redhat-marketplace-79mqm\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.079080 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.131854 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.148929 4751 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.148982 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.161857 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.162851 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.173678 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qg59s\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.175072 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.299537 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.322331 4751 generic.go:334] "Generic (PLEG): container finished" podID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerID="5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0" exitCode=0 Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.322383 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerDied","Data":"5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.325655 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" event={"ID":"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9","Type":"ContainerStarted","Data":"56a9589e4eab47cd80e7c6d15f8b62fc9480a78f9488abb398a2fb8b5b21528c"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.325698 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" event={"ID":"ca19e1d9-4d00-42dc-9754-c2f4e3d70ef9","Type":"ContainerStarted","Data":"4512a7ab606c2484693e2898ac332698addc6635fcbdfbd87595e1b29c9933b0"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.327705 4751 generic.go:334] "Generic (PLEG): container finished" podID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerID="4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7" exitCode=0 Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.327776 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerDied","Data":"4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.330608 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerID="da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4" exitCode=0 Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.330788 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerDied","Data":"da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.331477 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerStarted","Data":"6e93d08685adb75aa32339171f6214971eb078c1c820e9b69f9cd80a8ed94ab0"} Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.334645 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4prbp\" (UniqueName: \"kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.334834 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.335043 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.361573 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6swj5" podStartSLOduration=10.361556917 podStartE2EDuration="10.361556917s" podCreationTimestamp="2025-10-02 10:54:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:45.359473699 +0000 UTC m=+167.413700159" watchObservedRunningTime="2025-10-02 10:54:45.361556917 +0000 UTC m=+167.415783377" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.420904 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.436859 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.436971 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.437006 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4prbp\" (UniqueName: \"kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.438498 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.438702 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.461519 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4prbp\" (UniqueName: \"kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp\") pod \"redhat-marketplace-fwtkv\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.482652 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.539014 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.540036 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.544733 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.544917 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.547073 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.549922 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.583381 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.642778 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume\") pod \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.642883 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume\") pod \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.642957 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q442s\" (UniqueName: \"kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s\") pod \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\" (UID: \"3a8d6102-cb02-4184-9ad5-e1fa0a15b406\") " Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.643347 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.643431 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.643908 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume" (OuterVolumeSpecName: "config-volume") pod "3a8d6102-cb02-4184-9ad5-e1fa0a15b406" (UID: "3a8d6102-cb02-4184-9ad5-e1fa0a15b406"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.647824 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s" (OuterVolumeSpecName: "kube-api-access-q442s") pod "3a8d6102-cb02-4184-9ad5-e1fa0a15b406" (UID: "3a8d6102-cb02-4184-9ad5-e1fa0a15b406"). InnerVolumeSpecName "kube-api-access-q442s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.651258 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3a8d6102-cb02-4184-9ad5-e1fa0a15b406" (UID: "3a8d6102-cb02-4184-9ad5-e1fa0a15b406"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.733703 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:54:45 crc kubenswrapper[4751]: W1002 10:54:45.739128 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68293b1e_b00d_4757_9216_019326c55a63.slice/crio-14733a43e40f0658b3ce557b47a26ce3ccb562e158775e32c8e76141d43fcceb WatchSource:0}: Error finding container 14733a43e40f0658b3ce557b47a26ce3ccb562e158775e32c8e76141d43fcceb: Status 404 returned error can't find the container with id 14733a43e40f0658b3ce557b47a26ce3ccb562e158775e32c8e76141d43fcceb Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745704 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745760 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745833 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745847 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745857 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q442s\" (UniqueName: \"kubernetes.io/projected/3a8d6102-cb02-4184-9ad5-e1fa0a15b406-kube-api-access-q442s\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.745899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.763727 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.882617 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.909323 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:54:45 crc kubenswrapper[4751]: W1002 10:54:45.921684 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf27e1b4b_501e_4721_ba3e_c3b3d862e53c.slice/crio-03875ba66e8edc9604ea3ce614bfa91c437a303f58a13914beca02bd5c79d647 WatchSource:0}: Error finding container 03875ba66e8edc9604ea3ce614bfa91c437a303f58a13914beca02bd5c79d647: Status 404 returned error can't find the container with id 03875ba66e8edc9604ea3ce614bfa91c437a303f58a13914beca02bd5c79d647 Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.944202 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:45 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:45 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:45 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:45 crc kubenswrapper[4751]: I1002 10:54:45.944265 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.056528 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.165997 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:54:46 crc kubenswrapper[4751]: E1002 10:54:46.166204 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8d6102-cb02-4184-9ad5-e1fa0a15b406" containerName="collect-profiles" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.166215 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8d6102-cb02-4184-9ad5-e1fa0a15b406" containerName="collect-profiles" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.166308 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8d6102-cb02-4184-9ad5-e1fa0a15b406" containerName="collect-profiles" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.167621 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.169746 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.182787 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.343377 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" event={"ID":"3a8d6102-cb02-4184-9ad5-e1fa0a15b406","Type":"ContainerDied","Data":"7f26b159b38b46136e253da32471f30eee0a10b0f4c4d514fc19f14ededfc0c1"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.343688 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f26b159b38b46136e253da32471f30eee0a10b0f4c4d514fc19f14ededfc0c1" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.343467 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.345841 4751 generic.go:334] "Generic (PLEG): container finished" podID="68293b1e-b00d-4757-9216-019326c55a63" containerID="22cb0c12da44c544ca0a8ff84b13d6b30cebee9737ff439eec67bbbc99b04d9b" exitCode=0 Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.345915 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerDied","Data":"22cb0c12da44c544ca0a8ff84b13d6b30cebee9737ff439eec67bbbc99b04d9b"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.345939 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerStarted","Data":"14733a43e40f0658b3ce557b47a26ce3ccb562e158775e32c8e76141d43fcceb"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.347884 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a0e46a3-ab74-42fe-ae07-c67fd046e059","Type":"ContainerStarted","Data":"e79857c7e0313bef8baf3da120826db4161884eb21eb21fecf2376c827a1f751"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.353225 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" event={"ID":"f27e1b4b-501e-4721-ba3e-c3b3d862e53c","Type":"ContainerStarted","Data":"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.353289 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" event={"ID":"f27e1b4b-501e-4721-ba3e-c3b3d862e53c","Type":"ContainerStarted","Data":"03875ba66e8edc9604ea3ce614bfa91c437a303f58a13914beca02bd5c79d647"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.353310 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.354711 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5nwr\" (UniqueName: \"kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.354757 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.354776 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.355029 4751 generic.go:334] "Generic (PLEG): container finished" podID="3adadf57-57a6-4200-9890-6b3abcf24663" containerID="3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594" exitCode=0 Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.355303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerDied","Data":"3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.355368 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerStarted","Data":"738b3451dcb58fe40af7acf3c8e9088e8caeabc1c2facef35dfdbb0d96c27346"} Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.405592 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" podStartSLOduration=135.40557046200001 podStartE2EDuration="2m15.405570462s" podCreationTimestamp="2025-10-02 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:54:46.399226793 +0000 UTC m=+168.453453243" watchObservedRunningTime="2025-10-02 10:54:46.405570462 +0000 UTC m=+168.459796932" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.458117 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5nwr\" (UniqueName: \"kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.458328 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.458357 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.458957 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.459980 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.475685 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5nwr\" (UniqueName: \"kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr\") pod \"redhat-operators-mfdl4\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.558092 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.559046 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.572785 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.594385 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.660725 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwlx4\" (UniqueName: \"kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.660774 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.660809 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.763032 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.763387 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.763472 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwlx4\" (UniqueName: \"kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.763998 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.764038 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.786879 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwlx4\" (UniqueName: \"kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4\") pod \"redhat-operators-pmkqw\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.820502 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.878823 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.943276 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:46 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:46 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:46 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:46 crc kubenswrapper[4751]: I1002 10:54:46.943318 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.146715 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.288905 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.288944 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.291647 4751 patch_prober.go:28] interesting pod/console-f9d7485db-k2g7v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.291684 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k2g7v" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" probeResult="failure" output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.371042 4751 generic.go:334] "Generic (PLEG): container finished" podID="34f2464f-57f2-443e-8221-e21436dc9db0" containerID="59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569" exitCode=0 Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.371196 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerDied","Data":"59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569"} Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.371240 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerStarted","Data":"fc8d450a6496257d127e886378fd4160160364f5a2e44ae438d242dc7b4f20c8"} Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.376231 4751 generic.go:334] "Generic (PLEG): container finished" podID="3a0e46a3-ab74-42fe-ae07-c67fd046e059" containerID="7b6f4be3895c77babe979e2d7f1a958399c724e3f200b2446485b921f7871450" exitCode=0 Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.376303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a0e46a3-ab74-42fe-ae07-c67fd046e059","Type":"ContainerDied","Data":"7b6f4be3895c77babe979e2d7f1a958399c724e3f200b2446485b921f7871450"} Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.433024 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.437648 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-gg8df" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.461947 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.466153 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6qpgb" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.947631 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.948571 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:47 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:47 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:47 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.948635 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:47 crc kubenswrapper[4751]: I1002 10:54:47.991470 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5mlsz" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.183574 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.184002 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.184382 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.184450 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.314148 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.757050 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.757682 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.757743 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.760448 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.760638 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.898905 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.898949 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.943844 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:48 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:48 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:48 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:48 crc kubenswrapper[4751]: I1002 10:54:48.943922 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.000084 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.000209 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.000259 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.021040 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.089702 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.943069 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:49 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:49 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:49 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:49 crc kubenswrapper[4751]: I1002 10:54:49.943143 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:50 crc kubenswrapper[4751]: I1002 10:54:50.412876 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bwnql" Oct 02 10:54:50 crc kubenswrapper[4751]: I1002 10:54:50.943114 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:50 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:50 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:50 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:50 crc kubenswrapper[4751]: I1002 10:54:50.943249 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:51 crc kubenswrapper[4751]: I1002 10:54:51.943513 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:51 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:51 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:51 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:51 crc kubenswrapper[4751]: I1002 10:54:51.943558 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:52 crc kubenswrapper[4751]: I1002 10:54:52.943719 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:52 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:52 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:52 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:52 crc kubenswrapper[4751]: I1002 10:54:52.944097 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:53 crc kubenswrapper[4751]: I1002 10:54:53.583166 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:53 crc kubenswrapper[4751]: I1002 10:54:53.599020 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/278152b8-cf19-4211-95c9-2a162ff9549f-metrics-certs\") pod \"network-metrics-daemon-tzdtd\" (UID: \"278152b8-cf19-4211-95c9-2a162ff9549f\") " pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:53 crc kubenswrapper[4751]: I1002 10:54:53.867844 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzdtd" Oct 02 10:54:53 crc kubenswrapper[4751]: I1002 10:54:53.943921 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:53 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:53 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:53 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:53 crc kubenswrapper[4751]: I1002 10:54:53.943994 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:54 crc kubenswrapper[4751]: I1002 10:54:54.943936 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:54 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:54 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:54 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:54 crc kubenswrapper[4751]: I1002 10:54:54.944623 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:55 crc kubenswrapper[4751]: W1002 10:54:55.090394 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c41cf0a_041d_459b_ae0b_e112a4fef94f.slice/crio-b79ee41786102e2ecbfe683542548f28d47d116bf62a33c8855939f3870c14b6 WatchSource:0}: Error finding container b79ee41786102e2ecbfe683542548f28d47d116bf62a33c8855939f3870c14b6: Status 404 returned error can't find the container with id b79ee41786102e2ecbfe683542548f28d47d116bf62a33c8855939f3870c14b6 Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.159321 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.303953 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access\") pod \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.304738 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir\") pod \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\" (UID: \"3a0e46a3-ab74-42fe-ae07-c67fd046e059\") " Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.305125 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3a0e46a3-ab74-42fe-ae07-c67fd046e059" (UID: "3a0e46a3-ab74-42fe-ae07-c67fd046e059"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.305314 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.313621 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3a0e46a3-ab74-42fe-ae07-c67fd046e059" (UID: "3a0e46a3-ab74-42fe-ae07-c67fd046e059"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.406617 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a0e46a3-ab74-42fe-ae07-c67fd046e059-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.428859 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerStarted","Data":"b79ee41786102e2ecbfe683542548f28d47d116bf62a33c8855939f3870c14b6"} Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.430071 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3a0e46a3-ab74-42fe-ae07-c67fd046e059","Type":"ContainerDied","Data":"e79857c7e0313bef8baf3da120826db4161884eb21eb21fecf2376c827a1f751"} Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.430096 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79857c7e0313bef8baf3da120826db4161884eb21eb21fecf2376c827a1f751" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.430148 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.944074 4751 patch_prober.go:28] interesting pod/router-default-5444994796-jnthj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 10:54:55 crc kubenswrapper[4751]: [-]has-synced failed: reason withheld Oct 02 10:54:55 crc kubenswrapper[4751]: [+]process-running ok Oct 02 10:54:55 crc kubenswrapper[4751]: healthz check failed Oct 02 10:54:55 crc kubenswrapper[4751]: I1002 10:54:55.944156 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jnthj" podUID="9b35ed5d-5bc5-4df6-81f4-89e807ae69ed" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 10:54:56 crc kubenswrapper[4751]: I1002 10:54:56.943910 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:56 crc kubenswrapper[4751]: I1002 10:54:56.947221 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jnthj" Oct 02 10:54:57 crc kubenswrapper[4751]: I1002 10:54:57.289962 4751 patch_prober.go:28] interesting pod/console-f9d7485db-k2g7v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 02 10:54:57 crc kubenswrapper[4751]: I1002 10:54:57.290086 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-k2g7v" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" probeResult="failure" output="Get \"https://10.217.0.5:8443/health\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 02 10:54:58 crc kubenswrapper[4751]: I1002 10:54:58.182934 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:58 crc kubenswrapper[4751]: I1002 10:54:58.182985 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:54:58 crc kubenswrapper[4751]: I1002 10:54:58.183003 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:54:58 crc kubenswrapper[4751]: I1002 10:54:58.183043 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:01 crc kubenswrapper[4751]: I1002 10:55:01.506903 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:55:01 crc kubenswrapper[4751]: I1002 10:55:01.507783 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:55:04 crc kubenswrapper[4751]: I1002 10:55:04.040320 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 10:55:05 crc kubenswrapper[4751]: I1002 10:55:05.429739 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:55:07 crc kubenswrapper[4751]: I1002 10:55:07.345132 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:55:07 crc kubenswrapper[4751]: I1002 10:55:07.352535 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.183969 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.183973 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.184434 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.184489 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.184565 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.185123 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.185227 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.185422 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"8851fec37cab6527ab80b195d0586c1074d694be91a85e79d8d621d8f1412318"} pod="openshift-console/downloads-7954f5f757-2dt22" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 02 10:55:08 crc kubenswrapper[4751]: I1002 10:55:08.185563 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" containerID="cri-o://8851fec37cab6527ab80b195d0586c1074d694be91a85e79d8d621d8f1412318" gracePeriod=2 Oct 02 10:55:11 crc kubenswrapper[4751]: I1002 10:55:11.541373 4751 generic.go:334] "Generic (PLEG): container finished" podID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerID="8851fec37cab6527ab80b195d0586c1074d694be91a85e79d8d621d8f1412318" exitCode=0 Oct 02 10:55:11 crc kubenswrapper[4751]: I1002 10:55:11.541616 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2dt22" event={"ID":"f14b663f-94ad-44f2-82a8-ec6d53159146","Type":"ContainerDied","Data":"8851fec37cab6527ab80b195d0586c1074d694be91a85e79d8d621d8f1412318"} Oct 02 10:55:14 crc kubenswrapper[4751]: E1002 10:55:14.028397 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 10:55:14 crc kubenswrapper[4751]: E1002 10:55:14.028593 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tm26h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wjgj7_openshift-marketplace(fbf7ed8c-d710-4d79-bdfc-31a84398d154): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:14 crc kubenswrapper[4751]: E1002 10:55:14.029808 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wjgj7" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" Oct 02 10:55:15 crc kubenswrapper[4751]: E1002 10:55:15.162461 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wjgj7" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" Oct 02 10:55:15 crc kubenswrapper[4751]: I1002 10:55:15.570272 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 10:55:15 crc kubenswrapper[4751]: W1002 10:55:15.579429 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7777c51d_e76f_43a1_8133_5cc75b6f4a5d.slice/crio-eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c WatchSource:0}: Error finding container eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c: Status 404 returned error can't find the container with id eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c Oct 02 10:55:15 crc kubenswrapper[4751]: I1002 10:55:15.638216 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzdtd"] Oct 02 10:55:15 crc kubenswrapper[4751]: W1002 10:55:15.658281 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod278152b8_cf19_4211_95c9_2a162ff9549f.slice/crio-d0f0f19efc51e2d247bf8999db1846e42570a246c8ac969af33ef6d7945915e5 WatchSource:0}: Error finding container d0f0f19efc51e2d247bf8999db1846e42570a246c8ac969af33ef6d7945915e5: Status 404 returned error can't find the container with id d0f0f19efc51e2d247bf8999db1846e42570a246c8ac969af33ef6d7945915e5 Oct 02 10:55:16 crc kubenswrapper[4751]: E1002 10:55:16.252947 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 10:55:16 crc kubenswrapper[4751]: E1002 10:55:16.253584 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nfl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t46qb_openshift-marketplace(f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:16 crc kubenswrapper[4751]: E1002 10:55:16.254807 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t46qb" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" Oct 02 10:55:16 crc kubenswrapper[4751]: I1002 10:55:16.587226 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7777c51d-e76f-43a1-8133-5cc75b6f4a5d","Type":"ContainerStarted","Data":"eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c"} Oct 02 10:55:16 crc kubenswrapper[4751]: I1002 10:55:16.589253 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerID="7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74" exitCode=0 Oct 02 10:55:16 crc kubenswrapper[4751]: I1002 10:55:16.589309 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerDied","Data":"7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74"} Oct 02 10:55:16 crc kubenswrapper[4751]: I1002 10:55:16.593913 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" event={"ID":"278152b8-cf19-4211-95c9-2a162ff9549f","Type":"ContainerStarted","Data":"d0f0f19efc51e2d247bf8999db1846e42570a246c8ac969af33ef6d7945915e5"} Oct 02 10:55:16 crc kubenswrapper[4751]: E1002 10:55:16.596367 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t46qb" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" Oct 02 10:55:17 crc kubenswrapper[4751]: I1002 10:55:17.606912 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2dt22" event={"ID":"f14b663f-94ad-44f2-82a8-ec6d53159146","Type":"ContainerStarted","Data":"901280df26c427f78bb46f011572dc4e506706f2d498524c912c4006b71dc31f"} Oct 02 10:55:18 crc kubenswrapper[4751]: I1002 10:55:18.185289 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:18 crc kubenswrapper[4751]: I1002 10:55:18.185374 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:18 crc kubenswrapper[4751]: I1002 10:55:18.311217 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2m9f" Oct 02 10:55:19 crc kubenswrapper[4751]: I1002 10:55:19.621082 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" event={"ID":"278152b8-cf19-4211-95c9-2a162ff9549f","Type":"ContainerStarted","Data":"bbb126affd9eab151bb11c9098270d99408db3a0dc723cb164f7dc336638c103"} Oct 02 10:55:19 crc kubenswrapper[4751]: I1002 10:55:19.621350 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:55:19 crc kubenswrapper[4751]: I1002 10:55:19.621668 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:19 crc kubenswrapper[4751]: I1002 10:55:19.621714 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.235061 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.235406 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v9v7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nl7xm_openshift-marketplace(4257aeff-6e4a-4b1e-ae09-64065d3eaba8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.236688 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nl7xm" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.287426 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.287637 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgfdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-f26gx_openshift-marketplace(3e93c6df-97ab-4ac4-b130-c7a7127ca584): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.289064 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-f26gx" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" Oct 02 10:55:20 crc kubenswrapper[4751]: I1002 10:55:20.629450 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7777c51d-e76f-43a1-8133-5cc75b6f4a5d","Type":"ContainerStarted","Data":"15de68a34e8ba78ea3a75feed5fc54b62bdda5cc6ddef8771af3a46e14baff2a"} Oct 02 10:55:20 crc kubenswrapper[4751]: I1002 10:55:20.630276 4751 patch_prober.go:28] interesting pod/downloads-7954f5f757-2dt22 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 02 10:55:20 crc kubenswrapper[4751]: I1002 10:55:20.630366 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2dt22" podUID="f14b663f-94ad-44f2-82a8-ec6d53159146" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.632336 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-f26gx" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" Oct 02 10:55:20 crc kubenswrapper[4751]: E1002 10:55:20.633325 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nl7xm" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" Oct 02 10:55:21 crc kubenswrapper[4751]: I1002 10:55:21.658328 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=33.658296811 podStartE2EDuration="33.658296811s" podCreationTimestamp="2025-10-02 10:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:55:21.656782358 +0000 UTC m=+203.711008818" watchObservedRunningTime="2025-10-02 10:55:21.658296811 +0000 UTC m=+203.712523301" Oct 02 10:55:22 crc kubenswrapper[4751]: I1002 10:55:22.642830 4751 generic.go:334] "Generic (PLEG): container finished" podID="7777c51d-e76f-43a1-8133-5cc75b6f4a5d" containerID="15de68a34e8ba78ea3a75feed5fc54b62bdda5cc6ddef8771af3a46e14baff2a" exitCode=0 Oct 02 10:55:22 crc kubenswrapper[4751]: I1002 10:55:22.642890 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7777c51d-e76f-43a1-8133-5cc75b6f4a5d","Type":"ContainerDied","Data":"15de68a34e8ba78ea3a75feed5fc54b62bdda5cc6ddef8771af3a46e14baff2a"} Oct 02 10:55:22 crc kubenswrapper[4751]: I1002 10:55:22.647041 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzdtd" event={"ID":"278152b8-cf19-4211-95c9-2a162ff9549f","Type":"ContainerStarted","Data":"c01dbc3ffc8846395c19b2c2b4cf64255fcb1dc4d70edafd08aa51deb3db1389"} Oct 02 10:55:22 crc kubenswrapper[4751]: I1002 10:55:22.682777 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-tzdtd" podStartSLOduration=172.682759326 podStartE2EDuration="2m52.682759326s" podCreationTimestamp="2025-10-02 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:55:22.681972053 +0000 UTC m=+204.736198513" watchObservedRunningTime="2025-10-02 10:55:22.682759326 +0000 UTC m=+204.736985776" Oct 02 10:55:23 crc kubenswrapper[4751]: E1002 10:55:23.320868 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 10:55:23 crc kubenswrapper[4751]: E1002 10:55:23.321085 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4prbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fwtkv_openshift-marketplace(68293b1e-b00d-4757-9216-019326c55a63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\": context canceled" logger="UnhandledError" Oct 02 10:55:23 crc kubenswrapper[4751]: E1002 10:55:23.322281 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1: Get \\\"https://registry.redhat.io/v2/redhat/redhat-marketplace-index/blobs/sha256:4fcf0426ebc05442a4c6e577d2e4f80bebb28f88fd9b27d7c57520dcd918bed1\\\": context canceled\"" pod="openshift-marketplace/redhat-marketplace-fwtkv" podUID="68293b1e-b00d-4757-9216-019326c55a63" Oct 02 10:55:26 crc kubenswrapper[4751]: E1002 10:55:26.227578 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fwtkv" podUID="68293b1e-b00d-4757-9216-019326c55a63" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.272048 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.320809 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir\") pod \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.320875 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access\") pod \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\" (UID: \"7777c51d-e76f-43a1-8133-5cc75b6f4a5d\") " Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.321934 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7777c51d-e76f-43a1-8133-5cc75b6f4a5d" (UID: "7777c51d-e76f-43a1-8133-5cc75b6f4a5d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.341377 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7777c51d-e76f-43a1-8133-5cc75b6f4a5d" (UID: "7777c51d-e76f-43a1-8133-5cc75b6f4a5d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.422163 4751 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.422224 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7777c51d-e76f-43a1-8133-5cc75b6f4a5d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.671001 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.670952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7777c51d-e76f-43a1-8133-5cc75b6f4a5d","Type":"ContainerDied","Data":"eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c"} Oct 02 10:55:26 crc kubenswrapper[4751]: I1002 10:55:26.671141 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaac7ee2cf04ef664fcaaa80b0bd13112db7e26d2922065e0393541322619b4c" Oct 02 10:55:28 crc kubenswrapper[4751]: I1002 10:55:28.189954 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2dt22" Oct 02 10:55:29 crc kubenswrapper[4751]: E1002 10:55:29.404985 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 10:55:29 crc kubenswrapper[4751]: E1002 10:55:29.405209 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5nwr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mfdl4_openshift-marketplace(34f2464f-57f2-443e-8221-e21436dc9db0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:29 crc kubenswrapper[4751]: E1002 10:55:29.406397 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mfdl4" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" Oct 02 10:55:31 crc kubenswrapper[4751]: I1002 10:55:31.507971 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:55:31 crc kubenswrapper[4751]: I1002 10:55:31.508799 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:55:31 crc kubenswrapper[4751]: I1002 10:55:31.508938 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:55:31 crc kubenswrapper[4751]: I1002 10:55:31.510138 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:55:31 crc kubenswrapper[4751]: I1002 10:55:31.510290 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0" gracePeriod=600 Oct 02 10:55:32 crc kubenswrapper[4751]: I1002 10:55:32.722654 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0" exitCode=0 Oct 02 10:55:32 crc kubenswrapper[4751]: I1002 10:55:32.722763 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0"} Oct 02 10:55:35 crc kubenswrapper[4751]: E1002 10:55:35.219508 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mfdl4" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" Oct 02 10:55:56 crc kubenswrapper[4751]: E1002 10:55:56.254470 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 10:55:56 crc kubenswrapper[4751]: E1002 10:55:56.255863 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vwlx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pmkqw_openshift-marketplace(7c41cf0a-041d-459b-ae0b-e112a4fef94f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:55:56 crc kubenswrapper[4751]: E1002 10:55:56.257250 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pmkqw" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" Oct 02 10:55:56 crc kubenswrapper[4751]: E1002 10:55:56.886558 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pmkqw" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" Oct 02 10:55:59 crc kubenswrapper[4751]: I1002 10:55:59.905855 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49"} Oct 02 10:56:02 crc kubenswrapper[4751]: E1002 10:56:02.344589 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 10:56:02 crc kubenswrapper[4751]: E1002 10:56:02.346100 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gk47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-79mqm_openshift-marketplace(3adadf57-57a6-4200-9890-6b3abcf24663): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 10:56:02 crc kubenswrapper[4751]: E1002 10:56:02.347506 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-79mqm" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" Oct 02 10:56:17 crc kubenswrapper[4751]: I1002 10:56:17.028522 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerStarted","Data":"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca"} Oct 02 10:56:17 crc kubenswrapper[4751]: I1002 10:56:17.032141 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerStarted","Data":"fe5cafca3298a5c2fd1c7054f370cb517454ebd8cc7b214d5616d721e83d0102"} Oct 02 10:56:17 crc kubenswrapper[4751]: I1002 10:56:17.036636 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerStarted","Data":"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8"} Oct 02 10:56:17 crc kubenswrapper[4751]: I1002 10:56:17.038519 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerStarted","Data":"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5"} Oct 02 10:56:17 crc kubenswrapper[4751]: I1002 10:56:17.040915 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerStarted","Data":"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.061228 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerID="11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.061315 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerDied","Data":"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.064792 4751 generic.go:334] "Generic (PLEG): container finished" podID="68293b1e-b00d-4757-9216-019326c55a63" containerID="fe5cafca3298a5c2fd1c7054f370cb517454ebd8cc7b214d5616d721e83d0102" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.065085 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerDied","Data":"fe5cafca3298a5c2fd1c7054f370cb517454ebd8cc7b214d5616d721e83d0102"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.069196 4751 generic.go:334] "Generic (PLEG): container finished" podID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerID="e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.069245 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerDied","Data":"e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.071858 4751 generic.go:334] "Generic (PLEG): container finished" podID="34f2464f-57f2-443e-8221-e21436dc9db0" containerID="49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.071928 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerDied","Data":"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.074816 4751 generic.go:334] "Generic (PLEG): container finished" podID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerID="42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.074864 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerDied","Data":"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5"} Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.078809 4751 generic.go:334] "Generic (PLEG): container finished" podID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerID="03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff" exitCode=0 Oct 02 10:56:18 crc kubenswrapper[4751]: I1002 10:56:18.078868 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerDied","Data":"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff"} Oct 02 10:56:19 crc kubenswrapper[4751]: I1002 10:56:19.093185 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerID="b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482" exitCode=0 Oct 02 10:56:19 crc kubenswrapper[4751]: I1002 10:56:19.093266 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerDied","Data":"b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482"} Oct 02 10:56:19 crc kubenswrapper[4751]: I1002 10:56:19.096945 4751 generic.go:334] "Generic (PLEG): container finished" podID="3adadf57-57a6-4200-9890-6b3abcf24663" containerID="3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871" exitCode=0 Oct 02 10:56:19 crc kubenswrapper[4751]: I1002 10:56:19.096994 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerDied","Data":"3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871"} Oct 02 10:56:22 crc kubenswrapper[4751]: I1002 10:56:22.120202 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerStarted","Data":"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826"} Oct 02 10:56:22 crc kubenswrapper[4751]: I1002 10:56:22.143521 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f26gx" podStartSLOduration=3.5323035259999997 podStartE2EDuration="1m39.143490293s" podCreationTimestamp="2025-10-02 10:54:43 +0000 UTC" firstStartedPulling="2025-10-02 10:54:45.334098895 +0000 UTC m=+167.388325345" lastFinishedPulling="2025-10-02 10:56:20.945285662 +0000 UTC m=+262.999512112" observedRunningTime="2025-10-02 10:56:22.138734528 +0000 UTC m=+264.192960978" watchObservedRunningTime="2025-10-02 10:56:22.143490293 +0000 UTC m=+264.197716783" Oct 02 10:56:23 crc kubenswrapper[4751]: I1002 10:56:23.762067 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:23 crc kubenswrapper[4751]: I1002 10:56:23.763143 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:25 crc kubenswrapper[4751]: I1002 10:56:25.173747 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-f26gx" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="registry-server" probeResult="failure" output=< Oct 02 10:56:25 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 10:56:25 crc kubenswrapper[4751]: > Oct 02 10:56:30 crc kubenswrapper[4751]: I1002 10:56:30.180218 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerStarted","Data":"34d2454b384a697a3d942a38add45161f4d6bab2e3c4b3d814c673fa31986593"} Oct 02 10:56:30 crc kubenswrapper[4751]: I1002 10:56:30.198420 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fwtkv" podStartSLOduration=5.240808013 podStartE2EDuration="1m45.1983843s" podCreationTimestamp="2025-10-02 10:54:45 +0000 UTC" firstStartedPulling="2025-10-02 10:54:46.349486024 +0000 UTC m=+168.403712474" lastFinishedPulling="2025-10-02 10:56:26.307062271 +0000 UTC m=+268.361288761" observedRunningTime="2025-10-02 10:56:30.195884669 +0000 UTC m=+272.250111129" watchObservedRunningTime="2025-10-02 10:56:30.1983843 +0000 UTC m=+272.252610760" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.192665 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerStarted","Data":"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.197051 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerStarted","Data":"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.199312 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerStarted","Data":"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.201364 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerStarted","Data":"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.202952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerStarted","Data":"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.205303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerStarted","Data":"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888"} Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.217539 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t46qb" podStartSLOduration=3.76289141 podStartE2EDuration="1m48.217522395s" podCreationTimestamp="2025-10-02 10:54:43 +0000 UTC" firstStartedPulling="2025-10-02 10:54:45.329622859 +0000 UTC m=+167.383849309" lastFinishedPulling="2025-10-02 10:56:29.784253834 +0000 UTC m=+271.838480294" observedRunningTime="2025-10-02 10:56:31.214976693 +0000 UTC m=+273.269203143" watchObservedRunningTime="2025-10-02 10:56:31.217522395 +0000 UTC m=+273.271748855" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.236232 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wjgj7" podStartSLOduration=4.612749398 podStartE2EDuration="1m49.236209227s" podCreationTimestamp="2025-10-02 10:54:42 +0000 UTC" firstStartedPulling="2025-10-02 10:54:45.327357555 +0000 UTC m=+167.381584005" lastFinishedPulling="2025-10-02 10:56:29.950817374 +0000 UTC m=+272.005043834" observedRunningTime="2025-10-02 10:56:31.232187313 +0000 UTC m=+273.286413753" watchObservedRunningTime="2025-10-02 10:56:31.236209227 +0000 UTC m=+273.290435677" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.248787 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mfdl4" podStartSLOduration=10.36710968 podStartE2EDuration="1m45.248771345s" podCreationTimestamp="2025-10-02 10:54:46 +0000 UTC" firstStartedPulling="2025-10-02 10:54:55.073250456 +0000 UTC m=+177.127476916" lastFinishedPulling="2025-10-02 10:56:29.954912111 +0000 UTC m=+272.009138581" observedRunningTime="2025-10-02 10:56:31.246446119 +0000 UTC m=+273.300672599" watchObservedRunningTime="2025-10-02 10:56:31.248771345 +0000 UTC m=+273.302997805" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.265076 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pmkqw" podStartSLOduration=31.966242486 podStartE2EDuration="1m45.265054528s" podCreationTimestamp="2025-10-02 10:54:46 +0000 UTC" firstStartedPulling="2025-10-02 10:55:16.592251012 +0000 UTC m=+198.646477502" lastFinishedPulling="2025-10-02 10:56:29.891063074 +0000 UTC m=+271.945289544" observedRunningTime="2025-10-02 10:56:31.264005608 +0000 UTC m=+273.318232058" watchObservedRunningTime="2025-10-02 10:56:31.265054528 +0000 UTC m=+273.319280978" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.285843 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nl7xm" podStartSLOduration=3.633673696 podStartE2EDuration="1m49.285823439s" podCreationTimestamp="2025-10-02 10:54:42 +0000 UTC" firstStartedPulling="2025-10-02 10:54:44.302519691 +0000 UTC m=+166.356746141" lastFinishedPulling="2025-10-02 10:56:29.954669384 +0000 UTC m=+272.008895884" observedRunningTime="2025-10-02 10:56:31.28374028 +0000 UTC m=+273.337966730" watchObservedRunningTime="2025-10-02 10:56:31.285823439 +0000 UTC m=+273.340049899" Oct 02 10:56:31 crc kubenswrapper[4751]: I1002 10:56:31.303370 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-79mqm" podStartSLOduration=3.809500844 podStartE2EDuration="1m47.303351698s" podCreationTimestamp="2025-10-02 10:54:44 +0000 UTC" firstStartedPulling="2025-10-02 10:54:46.356439559 +0000 UTC m=+168.410666009" lastFinishedPulling="2025-10-02 10:56:29.850290413 +0000 UTC m=+271.904516863" observedRunningTime="2025-10-02 10:56:31.301416993 +0000 UTC m=+273.355643453" watchObservedRunningTime="2025-10-02 10:56:31.303351698 +0000 UTC m=+273.357578148" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.087958 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.088636 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.129829 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.278107 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.278306 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.325955 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.504447 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.504506 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.543755 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.806607 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:33 crc kubenswrapper[4751]: I1002 10:56:33.848246 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.079519 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.079593 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.126429 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.274601 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.277145 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.483421 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.483818 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:35 crc kubenswrapper[4751]: I1002 10:56:35.560207 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.274372 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.594760 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.594896 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.649515 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.879703 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.879749 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:36 crc kubenswrapper[4751]: I1002 10:56:36.915903 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:37 crc kubenswrapper[4751]: I1002 10:56:37.228878 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:56:37 crc kubenswrapper[4751]: I1002 10:56:37.229223 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f26gx" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="registry-server" containerID="cri-o://6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826" gracePeriod=2 Oct 02 10:56:37 crc kubenswrapper[4751]: I1002 10:56:37.301520 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:56:37 crc kubenswrapper[4751]: I1002 10:56:37.302774 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.198251 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.252121 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerID="6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826" exitCode=0 Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.252189 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerDied","Data":"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826"} Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.252232 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f26gx" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.252252 4751 scope.go:117] "RemoveContainer" containerID="6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.252231 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f26gx" event={"ID":"3e93c6df-97ab-4ac4-b130-c7a7127ca584","Type":"ContainerDied","Data":"6e93d08685adb75aa32339171f6214971eb078c1c820e9b69f9cd80a8ed94ab0"} Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.394666 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgfdj\" (UniqueName: \"kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj\") pod \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.395330 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities\") pod \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.395471 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content\") pod \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\" (UID: \"3e93c6df-97ab-4ac4-b130-c7a7127ca584\") " Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.396988 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities" (OuterVolumeSpecName: "utilities") pod "3e93c6df-97ab-4ac4-b130-c7a7127ca584" (UID: "3e93c6df-97ab-4ac4-b130-c7a7127ca584"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.401315 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj" (OuterVolumeSpecName: "kube-api-access-vgfdj") pod "3e93c6df-97ab-4ac4-b130-c7a7127ca584" (UID: "3e93c6df-97ab-4ac4-b130-c7a7127ca584"). InnerVolumeSpecName "kube-api-access-vgfdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.432763 4751 scope.go:117] "RemoveContainer" containerID="11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.447603 4751 scope.go:117] "RemoveContainer" containerID="da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.468783 4751 scope.go:117] "RemoveContainer" containerID="6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826" Oct 02 10:56:40 crc kubenswrapper[4751]: E1002 10:56:39.469209 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826\": container with ID starting with 6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826 not found: ID does not exist" containerID="6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.469245 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826"} err="failed to get container status \"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826\": rpc error: code = NotFound desc = could not find container \"6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826\": container with ID starting with 6fdf7414ab2920ae045ac27402e38f3dccfc0b71993d270927a5b01e7dcb2826 not found: ID does not exist" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.469270 4751 scope.go:117] "RemoveContainer" containerID="11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca" Oct 02 10:56:40 crc kubenswrapper[4751]: E1002 10:56:39.469790 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca\": container with ID starting with 11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca not found: ID does not exist" containerID="11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.469839 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca"} err="failed to get container status \"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca\": rpc error: code = NotFound desc = could not find container \"11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca\": container with ID starting with 11288f68fefc41cb193c5c56f322e97e4c30b0b5ac91ea29ffd76a30567f41ca not found: ID does not exist" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.469875 4751 scope.go:117] "RemoveContainer" containerID="da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4" Oct 02 10:56:40 crc kubenswrapper[4751]: E1002 10:56:39.470341 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4\": container with ID starting with da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4 not found: ID does not exist" containerID="da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.470425 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4"} err="failed to get container status \"da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4\": rpc error: code = NotFound desc = could not find container \"da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4\": container with ID starting with da905f14974e62b73b4f7d967cedf3e34ceb4c7b0bdfe66ca672534df6d914f4 not found: ID does not exist" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.496830 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.496848 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgfdj\" (UniqueName: \"kubernetes.io/projected/3e93c6df-97ab-4ac4-b130-c7a7127ca584-kube-api-access-vgfdj\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.628638 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.628826 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fwtkv" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="registry-server" containerID="cri-o://34d2454b384a697a3d942a38add45161f4d6bab2e3c4b3d814c673fa31986593" gracePeriod=2 Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.829767 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.830160 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pmkqw" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="registry-server" containerID="cri-o://08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1" gracePeriod=2 Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:39.957462 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e93c6df-97ab-4ac4-b130-c7a7127ca584" (UID: "3e93c6df-97ab-4ac4-b130-c7a7127ca584"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:40.004351 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e93c6df-97ab-4ac4-b130-c7a7127ca584-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:40.187163 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:56:40 crc kubenswrapper[4751]: I1002 10:56:40.190941 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f26gx"] Oct 02 10:56:41 crc kubenswrapper[4751]: I1002 10:56:41.562090 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" path="/var/lib/kubelet/pods/3e93c6df-97ab-4ac4-b130-c7a7127ca584/volumes" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.029931 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.132981 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content\") pod \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.133014 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities\") pod \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.133042 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwlx4\" (UniqueName: \"kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4\") pod \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\" (UID: \"7c41cf0a-041d-459b-ae0b-e112a4fef94f\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.137480 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities" (OuterVolumeSpecName: "utilities") pod "7c41cf0a-041d-459b-ae0b-e112a4fef94f" (UID: "7c41cf0a-041d-459b-ae0b-e112a4fef94f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.140340 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4" (OuterVolumeSpecName: "kube-api-access-vwlx4") pod "7c41cf0a-041d-459b-ae0b-e112a4fef94f" (UID: "7c41cf0a-041d-459b-ae0b-e112a4fef94f"). InnerVolumeSpecName "kube-api-access-vwlx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.234726 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.234792 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwlx4\" (UniqueName: \"kubernetes.io/projected/7c41cf0a-041d-459b-ae0b-e112a4fef94f-kube-api-access-vwlx4\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.243668 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c41cf0a-041d-459b-ae0b-e112a4fef94f" (UID: "7c41cf0a-041d-459b-ae0b-e112a4fef94f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.273270 4751 generic.go:334] "Generic (PLEG): container finished" podID="68293b1e-b00d-4757-9216-019326c55a63" containerID="34d2454b384a697a3d942a38add45161f4d6bab2e3c4b3d814c673fa31986593" exitCode=0 Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.273340 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerDied","Data":"34d2454b384a697a3d942a38add45161f4d6bab2e3c4b3d814c673fa31986593"} Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.276473 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerID="08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1" exitCode=0 Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.276504 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerDied","Data":"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1"} Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.276520 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pmkqw" event={"ID":"7c41cf0a-041d-459b-ae0b-e112a4fef94f","Type":"ContainerDied","Data":"b79ee41786102e2ecbfe683542548f28d47d116bf62a33c8855939f3870c14b6"} Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.276537 4751 scope.go:117] "RemoveContainer" containerID="08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.276639 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pmkqw" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.311600 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.313357 4751 scope.go:117] "RemoveContainer" containerID="b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.319557 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pmkqw"] Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.335147 4751 scope.go:117] "RemoveContainer" containerID="7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.335660 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c41cf0a-041d-459b-ae0b-e112a4fef94f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.368070 4751 scope.go:117] "RemoveContainer" containerID="08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1" Oct 02 10:56:42 crc kubenswrapper[4751]: E1002 10:56:42.368735 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1\": container with ID starting with 08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1 not found: ID does not exist" containerID="08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.368776 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1"} err="failed to get container status \"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1\": rpc error: code = NotFound desc = could not find container \"08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1\": container with ID starting with 08208c69ba9e76ebd0a2dafc7388e56a08a183a5490f94cba6c320b6785930d1 not found: ID does not exist" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.368803 4751 scope.go:117] "RemoveContainer" containerID="b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482" Oct 02 10:56:42 crc kubenswrapper[4751]: E1002 10:56:42.369257 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482\": container with ID starting with b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482 not found: ID does not exist" containerID="b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.369300 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482"} err="failed to get container status \"b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482\": rpc error: code = NotFound desc = could not find container \"b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482\": container with ID starting with b9233818ae63d09c094bb428cd3d57994b55cbadb0d7f748bfd5acdb3dc28482 not found: ID does not exist" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.369323 4751 scope.go:117] "RemoveContainer" containerID="7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74" Oct 02 10:56:42 crc kubenswrapper[4751]: E1002 10:56:42.369599 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74\": container with ID starting with 7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74 not found: ID does not exist" containerID="7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.369674 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74"} err="failed to get container status \"7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74\": rpc error: code = NotFound desc = could not find container \"7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74\": container with ID starting with 7d97ccb202b7ab2f3fd752bc762171c2248f836a71d8d212d05364516b435c74 not found: ID does not exist" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.492747 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.639835 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4prbp\" (UniqueName: \"kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp\") pod \"68293b1e-b00d-4757-9216-019326c55a63\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.639904 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content\") pod \"68293b1e-b00d-4757-9216-019326c55a63\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.640048 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities\") pod \"68293b1e-b00d-4757-9216-019326c55a63\" (UID: \"68293b1e-b00d-4757-9216-019326c55a63\") " Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.641849 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities" (OuterVolumeSpecName: "utilities") pod "68293b1e-b00d-4757-9216-019326c55a63" (UID: "68293b1e-b00d-4757-9216-019326c55a63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.645404 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp" (OuterVolumeSpecName: "kube-api-access-4prbp") pod "68293b1e-b00d-4757-9216-019326c55a63" (UID: "68293b1e-b00d-4757-9216-019326c55a63"). InnerVolumeSpecName "kube-api-access-4prbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.660515 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68293b1e-b00d-4757-9216-019326c55a63" (UID: "68293b1e-b00d-4757-9216-019326c55a63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.741336 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4prbp\" (UniqueName: \"kubernetes.io/projected/68293b1e-b00d-4757-9216-019326c55a63-kube-api-access-4prbp\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.741376 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:42 crc kubenswrapper[4751]: I1002 10:56:42.741388 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68293b1e-b00d-4757-9216-019326c55a63-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.189230 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.287097 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwtkv" event={"ID":"68293b1e-b00d-4757-9216-019326c55a63","Type":"ContainerDied","Data":"14733a43e40f0658b3ce557b47a26ce3ccb562e158775e32c8e76141d43fcceb"} Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.287227 4751 scope.go:117] "RemoveContainer" containerID="34d2454b384a697a3d942a38add45161f4d6bab2e3c4b3d814c673fa31986593" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.287129 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwtkv" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.317396 4751 scope.go:117] "RemoveContainer" containerID="fe5cafca3298a5c2fd1c7054f370cb517454ebd8cc7b214d5616d721e83d0102" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.337311 4751 scope.go:117] "RemoveContainer" containerID="22cb0c12da44c544ca0a8ff84b13d6b30cebee9737ff439eec67bbbc99b04d9b" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.347153 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.349616 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwtkv"] Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.540012 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.556947 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68293b1e-b00d-4757-9216-019326c55a63" path="/var/lib/kubelet/pods/68293b1e-b00d-4757-9216-019326c55a63/volumes" Oct 02 10:56:43 crc kubenswrapper[4751]: I1002 10:56:43.557976 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" path="/var/lib/kubelet/pods/7c41cf0a-041d-459b-ae0b-e112a4fef94f/volumes" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.228483 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.230146 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t46qb" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="registry-server" containerID="cri-o://8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823" gracePeriod=2 Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.680635 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.797630 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities\") pod \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.797678 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content\") pod \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.797703 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nfl9\" (UniqueName: \"kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9\") pod \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\" (UID: \"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc\") " Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.799853 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities" (OuterVolumeSpecName: "utilities") pod "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" (UID: "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.804113 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9" (OuterVolumeSpecName: "kube-api-access-8nfl9") pod "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" (UID: "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc"). InnerVolumeSpecName "kube-api-access-8nfl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.854608 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" (UID: "f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.899233 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.899295 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:46 crc kubenswrapper[4751]: I1002 10:56:46.899319 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nfl9\" (UniqueName: \"kubernetes.io/projected/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc-kube-api-access-8nfl9\") on node \"crc\" DevicePath \"\"" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.320953 4751 generic.go:334] "Generic (PLEG): container finished" podID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerID="8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823" exitCode=0 Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.321021 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerDied","Data":"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823"} Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.321064 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t46qb" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.321088 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t46qb" event={"ID":"f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc","Type":"ContainerDied","Data":"3cc62933d6c3dfa72573cdc0f1cd54f177f4175a712bc339a5075d9fda783c65"} Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.321113 4751 scope.go:117] "RemoveContainer" containerID="8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.342855 4751 scope.go:117] "RemoveContainer" containerID="03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.365112 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.369404 4751 scope.go:117] "RemoveContainer" containerID="4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.372432 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t46qb"] Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.391743 4751 scope.go:117] "RemoveContainer" containerID="8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823" Oct 02 10:56:47 crc kubenswrapper[4751]: E1002 10:56:47.392353 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823\": container with ID starting with 8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823 not found: ID does not exist" containerID="8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.392409 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823"} err="failed to get container status \"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823\": rpc error: code = NotFound desc = could not find container \"8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823\": container with ID starting with 8e7742222ff2d3684554a5accf18add68060d16623acf2de81d397c2492c2823 not found: ID does not exist" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.392446 4751 scope.go:117] "RemoveContainer" containerID="03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff" Oct 02 10:56:47 crc kubenswrapper[4751]: E1002 10:56:47.392749 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff\": container with ID starting with 03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff not found: ID does not exist" containerID="03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.392775 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff"} err="failed to get container status \"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff\": rpc error: code = NotFound desc = could not find container \"03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff\": container with ID starting with 03e70e57ce1e56b41b64d969295293dd9491837239732332e7cb29abb147c7ff not found: ID does not exist" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.392793 4751 scope.go:117] "RemoveContainer" containerID="4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7" Oct 02 10:56:47 crc kubenswrapper[4751]: E1002 10:56:47.393042 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7\": container with ID starting with 4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7 not found: ID does not exist" containerID="4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.393067 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7"} err="failed to get container status \"4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7\": rpc error: code = NotFound desc = could not find container \"4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7\": container with ID starting with 4dc260d2f2fcaff73a15ace01df7ce806010e658b2a0ffefb923b36d0e0e15f7 not found: ID does not exist" Oct 02 10:56:47 crc kubenswrapper[4751]: I1002 10:56:47.561841 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" path="/var/lib/kubelet/pods/f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc/volumes" Oct 02 10:57:28 crc kubenswrapper[4751]: I1002 10:57:28.606414 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:57:53 crc kubenswrapper[4751]: I1002 10:57:53.639224 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerName="oauth-openshift" containerID="cri-o://e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2" gracePeriod=15 Oct 02 10:57:53 crc kubenswrapper[4751]: E1002 10:57:53.766415 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad4df9d0_8e5f_4e23_807b_6ebebcc5433a.slice/crio-e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.591093 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.637937 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-85766c7959-4nkxr"] Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.638954 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerName="oauth-openshift" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.639209 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerName="oauth-openshift" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.639403 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.639581 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.639729 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.639842 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.639985 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.640102 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.640325 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7777c51d-e76f-43a1-8133-5cc75b6f4a5d" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.641014 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7777c51d-e76f-43a1-8133-5cc75b6f4a5d" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.641152 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.641319 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.643415 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.643576 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.643711 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.643826 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.643934 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.644039 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.645082 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.645379 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.645503 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.645623 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.645747 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.645855 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.645959 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0e46a3-ab74-42fe-ae07-c67fd046e059" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.646049 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0e46a3-ab74-42fe-ae07-c67fd046e059" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.646161 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.646313 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="extract-utilities" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.646433 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.646511 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="extract-content" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.646825 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c41cf0a-041d-459b-ae0b-e112a4fef94f" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.646922 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0da6eb8-21b5-4e3a-8bdb-2c86c70c4bbc" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.647008 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7777c51d-e76f-43a1-8133-5cc75b6f4a5d" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.647097 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e93c6df-97ab-4ac4-b130-c7a7127ca584" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.647213 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerName="oauth-openshift" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.647308 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="68293b1e-b00d-4757-9216-019326c55a63" containerName="registry-server" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.647408 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0e46a3-ab74-42fe-ae07-c67fd046e059" containerName="pruner" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.648239 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.648746 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85766c7959-4nkxr"] Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.681504 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.681900 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682004 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682034 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682321 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682417 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682450 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682476 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682498 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682528 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682572 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682600 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tktsq\" (UniqueName: \"kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682623 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682695 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.682712 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies\") pod \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\" (UID: \"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a\") " Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.683163 4751 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.684111 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.684294 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.684564 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.684572 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.689029 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.689939 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.690269 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.690770 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.691152 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq" (OuterVolumeSpecName: "kube-api-access-tktsq") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "kube-api-access-tktsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.691376 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.691579 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.692441 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.692784 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" (UID: "ad4df9d0-8e5f-4e23-807b-6ebebcc5433a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.728701 4751 generic.go:334] "Generic (PLEG): container finished" podID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" containerID="e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2" exitCode=0 Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.728787 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.728784 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" event={"ID":"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a","Type":"ContainerDied","Data":"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2"} Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.728972 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ztmtr" event={"ID":"ad4df9d0-8e5f-4e23-807b-6ebebcc5433a","Type":"ContainerDied","Data":"54d9c7286ed283f9cb8050e46d4ef9626d0aa3c80ecaf04be2249f1f62944149"} Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.729038 4751 scope.go:117] "RemoveContainer" containerID="e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.758076 4751 scope.go:117] "RemoveContainer" containerID="e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2" Oct 02 10:57:54 crc kubenswrapper[4751]: E1002 10:57:54.758919 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2\": container with ID starting with e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2 not found: ID does not exist" containerID="e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.758978 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2"} err="failed to get container status \"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2\": rpc error: code = NotFound desc = could not find container \"e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2\": container with ID starting with e177c76cb2148e3e4885c4daa32a4e8ac04bae84ae8ef9595f3de56e98bdc1a2 not found: ID does not exist" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.766869 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.771666 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ztmtr"] Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.784323 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.784727 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-session\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.784907 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.785071 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-service-ca\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.785367 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-router-certs\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.785618 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.785994 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.786203 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.786438 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-login\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.786657 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-dir\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.786830 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-policies\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.787009 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.787291 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j52ll\" (UniqueName: \"kubernetes.io/projected/1ef27e65-4dec-4d18-b275-99bdc73c4196-kube-api-access-j52ll\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.787497 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-error\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.787753 4751 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.787939 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.788104 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.788454 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.788663 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.788798 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.788937 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.789058 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.789254 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.789508 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.789730 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.789861 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tktsq\" (UniqueName: \"kubernetes.io/projected/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-kube-api-access-tktsq\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.790042 4751 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.891840 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.892404 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-session\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.892587 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-service-ca\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.892775 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.892984 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-router-certs\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893113 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893268 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893393 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893492 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-login\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893590 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-dir\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893713 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-policies\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893834 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893943 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j52ll\" (UniqueName: \"kubernetes.io/projected/1ef27e65-4dec-4d18-b275-99bdc73c4196-kube-api-access-j52ll\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.894033 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.894066 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-error\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.893843 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-service-ca\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.894492 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-dir\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.895050 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.895579 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1ef27e65-4dec-4d18-b275-99bdc73c4196-audit-policies\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.896992 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-session\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.897590 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.898564 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.900112 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.900706 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-login\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.900797 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-system-router-certs\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.901088 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.901329 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1ef27e65-4dec-4d18-b275-99bdc73c4196-v4-0-config-user-template-error\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.919346 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j52ll\" (UniqueName: \"kubernetes.io/projected/1ef27e65-4dec-4d18-b275-99bdc73c4196-kube-api-access-j52ll\") pod \"oauth-openshift-85766c7959-4nkxr\" (UID: \"1ef27e65-4dec-4d18-b275-99bdc73c4196\") " pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:54 crc kubenswrapper[4751]: I1002 10:57:54.969201 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.222815 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85766c7959-4nkxr"] Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.559773 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4df9d0-8e5f-4e23-807b-6ebebcc5433a" path="/var/lib/kubelet/pods/ad4df9d0-8e5f-4e23-807b-6ebebcc5433a/volumes" Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.740192 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" event={"ID":"1ef27e65-4dec-4d18-b275-99bdc73c4196","Type":"ContainerStarted","Data":"944c6bab1e799629962b258c2e20be2e128789e903acb32118ccc44a23b90079"} Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.740261 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" event={"ID":"1ef27e65-4dec-4d18-b275-99bdc73c4196","Type":"ContainerStarted","Data":"2ff15fa468a7b8aacb96e8635f02dcf385aea0660028ecde38826bf79e179287"} Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.740721 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:57:55 crc kubenswrapper[4751]: I1002 10:57:55.772467 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" podStartSLOduration=27.772438983 podStartE2EDuration="27.772438983s" podCreationTimestamp="2025-10-02 10:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:57:55.770215667 +0000 UTC m=+357.824442137" watchObservedRunningTime="2025-10-02 10:57:55.772438983 +0000 UTC m=+357.826665473" Oct 02 10:57:56 crc kubenswrapper[4751]: I1002 10:57:56.072583 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-85766c7959-4nkxr" Oct 02 10:58:01 crc kubenswrapper[4751]: I1002 10:58:01.507342 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:58:01 crc kubenswrapper[4751]: I1002 10:58:01.508078 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:58:31 crc kubenswrapper[4751]: I1002 10:58:31.507557 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:58:31 crc kubenswrapper[4751]: I1002 10:58:31.508317 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.054585 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.054915 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wjgj7" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="registry-server" containerID="cri-o://814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c" gracePeriod=30 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.067975 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.068350 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nl7xm" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="registry-server" containerID="cri-o://e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d" gracePeriod=30 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.079564 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.079787 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" containerID="cri-o://aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353" gracePeriod=30 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.092734 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.093102 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-79mqm" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="registry-server" containerID="cri-o://e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888" gracePeriod=30 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.098765 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.099098 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mfdl4" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="registry-server" containerID="cri-o://fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14" gracePeriod=30 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.116102 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vdm72"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.116982 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.131957 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vdm72"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.193298 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94dce789-2750-40a1-b622-11ad58a438df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.193420 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtpz\" (UniqueName: \"kubernetes.io/projected/94dce789-2750-40a1-b622-11ad58a438df-kube-api-access-lhtpz\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.193452 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94dce789-2750-40a1-b622-11ad58a438df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.293894 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtpz\" (UniqueName: \"kubernetes.io/projected/94dce789-2750-40a1-b622-11ad58a438df-kube-api-access-lhtpz\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.293945 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94dce789-2750-40a1-b622-11ad58a438df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.294001 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94dce789-2750-40a1-b622-11ad58a438df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.295648 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94dce789-2750-40a1-b622-11ad58a438df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.299241 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/94dce789-2750-40a1-b622-11ad58a438df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.311689 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtpz\" (UniqueName: \"kubernetes.io/projected/94dce789-2750-40a1-b622-11ad58a438df-kube-api-access-lhtpz\") pod \"marketplace-operator-79b997595-vdm72\" (UID: \"94dce789-2750-40a1-b622-11ad58a438df\") " pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.442500 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.553271 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.568008 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.568143 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.585515 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.589736 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703553 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content\") pod \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703619 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm26h\" (UniqueName: \"kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h\") pod \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703644 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities\") pod \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\" (UID: \"fbf7ed8c-d710-4d79-bdfc-31a84398d154\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703681 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gk47\" (UniqueName: \"kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47\") pod \"3adadf57-57a6-4200-9890-6b3abcf24663\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703701 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities\") pod \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703728 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca\") pod \"4718f397-506a-4690-8942-caecf3a3e6c6\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703762 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities\") pod \"3adadf57-57a6-4200-9890-6b3abcf24663\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703803 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities\") pod \"34f2464f-57f2-443e-8221-e21436dc9db0\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703830 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content\") pod \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703855 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics\") pod \"4718f397-506a-4690-8942-caecf3a3e6c6\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703877 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content\") pod \"3adadf57-57a6-4200-9890-6b3abcf24663\" (UID: \"3adadf57-57a6-4200-9890-6b3abcf24663\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703893 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w4wn\" (UniqueName: \"kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn\") pod \"4718f397-506a-4690-8942-caecf3a3e6c6\" (UID: \"4718f397-506a-4690-8942-caecf3a3e6c6\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703940 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9v7v\" (UniqueName: \"kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v\") pod \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\" (UID: \"4257aeff-6e4a-4b1e-ae09-64065d3eaba8\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703958 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5nwr\" (UniqueName: \"kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr\") pod \"34f2464f-57f2-443e-8221-e21436dc9db0\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.703975 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content\") pod \"34f2464f-57f2-443e-8221-e21436dc9db0\" (UID: \"34f2464f-57f2-443e-8221-e21436dc9db0\") " Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.705670 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities" (OuterVolumeSpecName: "utilities") pod "4257aeff-6e4a-4b1e-ae09-64065d3eaba8" (UID: "4257aeff-6e4a-4b1e-ae09-64065d3eaba8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.706127 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities" (OuterVolumeSpecName: "utilities") pod "34f2464f-57f2-443e-8221-e21436dc9db0" (UID: "34f2464f-57f2-443e-8221-e21436dc9db0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.706300 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities" (OuterVolumeSpecName: "utilities") pod "3adadf57-57a6-4200-9890-6b3abcf24663" (UID: "3adadf57-57a6-4200-9890-6b3abcf24663"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.706382 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities" (OuterVolumeSpecName: "utilities") pod "fbf7ed8c-d710-4d79-bdfc-31a84398d154" (UID: "fbf7ed8c-d710-4d79-bdfc-31a84398d154"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.709530 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h" (OuterVolumeSpecName: "kube-api-access-tm26h") pod "fbf7ed8c-d710-4d79-bdfc-31a84398d154" (UID: "fbf7ed8c-d710-4d79-bdfc-31a84398d154"). InnerVolumeSpecName "kube-api-access-tm26h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.709601 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4718f397-506a-4690-8942-caecf3a3e6c6" (UID: "4718f397-506a-4690-8942-caecf3a3e6c6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.709753 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47" (OuterVolumeSpecName: "kube-api-access-9gk47") pod "3adadf57-57a6-4200-9890-6b3abcf24663" (UID: "3adadf57-57a6-4200-9890-6b3abcf24663"). InnerVolumeSpecName "kube-api-access-9gk47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.710522 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v" (OuterVolumeSpecName: "kube-api-access-v9v7v") pod "4257aeff-6e4a-4b1e-ae09-64065d3eaba8" (UID: "4257aeff-6e4a-4b1e-ae09-64065d3eaba8"). InnerVolumeSpecName "kube-api-access-v9v7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.717065 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr" (OuterVolumeSpecName: "kube-api-access-g5nwr") pod "34f2464f-57f2-443e-8221-e21436dc9db0" (UID: "34f2464f-57f2-443e-8221-e21436dc9db0"). InnerVolumeSpecName "kube-api-access-g5nwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.717515 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4718f397-506a-4690-8942-caecf3a3e6c6" (UID: "4718f397-506a-4690-8942-caecf3a3e6c6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.718085 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn" (OuterVolumeSpecName: "kube-api-access-2w4wn") pod "4718f397-506a-4690-8942-caecf3a3e6c6" (UID: "4718f397-506a-4690-8942-caecf3a3e6c6"). InnerVolumeSpecName "kube-api-access-2w4wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.721655 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3adadf57-57a6-4200-9890-6b3abcf24663" (UID: "3adadf57-57a6-4200-9890-6b3abcf24663"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.760241 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbf7ed8c-d710-4d79-bdfc-31a84398d154" (UID: "fbf7ed8c-d710-4d79-bdfc-31a84398d154"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.763900 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4257aeff-6e4a-4b1e-ae09-64065d3eaba8" (UID: "4257aeff-6e4a-4b1e-ae09-64065d3eaba8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.789786 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34f2464f-57f2-443e-8221-e21436dc9db0" (UID: "34f2464f-57f2-443e-8221-e21436dc9db0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.805934 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9v7v\" (UniqueName: \"kubernetes.io/projected/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-kube-api-access-v9v7v\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806052 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5nwr\" (UniqueName: \"kubernetes.io/projected/34f2464f-57f2-443e-8221-e21436dc9db0-kube-api-access-g5nwr\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806115 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806217 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806310 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm26h\" (UniqueName: \"kubernetes.io/projected/fbf7ed8c-d710-4d79-bdfc-31a84398d154-kube-api-access-tm26h\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806375 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbf7ed8c-d710-4d79-bdfc-31a84398d154-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806433 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gk47\" (UniqueName: \"kubernetes.io/projected/3adadf57-57a6-4200-9890-6b3abcf24663-kube-api-access-9gk47\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806496 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806552 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806617 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806674 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f2464f-57f2-443e-8221-e21436dc9db0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806727 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4257aeff-6e4a-4b1e-ae09-64065d3eaba8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806790 4751 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4718f397-506a-4690-8942-caecf3a3e6c6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806848 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3adadf57-57a6-4200-9890-6b3abcf24663-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.806907 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w4wn\" (UniqueName: \"kubernetes.io/projected/4718f397-506a-4690-8942-caecf3a3e6c6-kube-api-access-2w4wn\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.845186 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vdm72"] Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.965002 4751 generic.go:334] "Generic (PLEG): container finished" podID="4718f397-506a-4690-8942-caecf3a3e6c6" containerID="aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353" exitCode=0 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.965106 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" event={"ID":"4718f397-506a-4690-8942-caecf3a3e6c6","Type":"ContainerDied","Data":"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.965522 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" event={"ID":"4718f397-506a-4690-8942-caecf3a3e6c6","Type":"ContainerDied","Data":"e152a3e2390d495eb0b65a270c27a973455296e764447a780aa22761d64e65f3"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.965154 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ht64t" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.965559 4751 scope.go:117] "RemoveContainer" containerID="aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.966896 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" event={"ID":"94dce789-2750-40a1-b622-11ad58a438df","Type":"ContainerStarted","Data":"b3b0ba827b7d8f94666e0446ece131ac8a87e7232b6c91dec32fa172c27c9cf0"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.969428 4751 generic.go:334] "Generic (PLEG): container finished" podID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerID="e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d" exitCode=0 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.969479 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerDied","Data":"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.969505 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nl7xm" event={"ID":"4257aeff-6e4a-4b1e-ae09-64065d3eaba8","Type":"ContainerDied","Data":"8a9eab9681b4fe9f876425da27391da1bde1d42b2f998a8814fff987e4cc160c"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.969579 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nl7xm" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.974212 4751 generic.go:334] "Generic (PLEG): container finished" podID="34f2464f-57f2-443e-8221-e21436dc9db0" containerID="fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14" exitCode=0 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.974303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerDied","Data":"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.974307 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfdl4" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.974322 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfdl4" event={"ID":"34f2464f-57f2-443e-8221-e21436dc9db0","Type":"ContainerDied","Data":"fc8d450a6496257d127e886378fd4160160364f5a2e44ae438d242dc7b4f20c8"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.976699 4751 generic.go:334] "Generic (PLEG): container finished" podID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerID="814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c" exitCode=0 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.976750 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerDied","Data":"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.976765 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wjgj7" event={"ID":"fbf7ed8c-d710-4d79-bdfc-31a84398d154","Type":"ContainerDied","Data":"e9a808cb4fa72310d707f9601f1c1a4b92a9d6661000a98d3919272a938e06aa"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.976850 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wjgj7" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.981151 4751 generic.go:334] "Generic (PLEG): container finished" podID="3adadf57-57a6-4200-9890-6b3abcf24663" containerID="e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888" exitCode=0 Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.981211 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerDied","Data":"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.981247 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79mqm" event={"ID":"3adadf57-57a6-4200-9890-6b3abcf24663","Type":"ContainerDied","Data":"738b3451dcb58fe40af7acf3c8e9088e8caeabc1c2facef35dfdbb0d96c27346"} Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.981332 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79mqm" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.987280 4751 scope.go:117] "RemoveContainer" containerID="aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353" Oct 02 10:58:32 crc kubenswrapper[4751]: E1002 10:58:32.987916 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353\": container with ID starting with aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353 not found: ID does not exist" containerID="aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.987988 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353"} err="failed to get container status \"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353\": rpc error: code = NotFound desc = could not find container \"aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353\": container with ID starting with aa7665e8b7e9ad257013dbaa9f3d76b20f35e308185e0d0b8ad1bf55e7d18353 not found: ID does not exist" Oct 02 10:58:32 crc kubenswrapper[4751]: I1002 10:58:32.988040 4751 scope.go:117] "RemoveContainer" containerID="e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.007371 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.014271 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ht64t"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.017269 4751 scope.go:117] "RemoveContainer" containerID="e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.020047 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.020083 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nl7xm"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.031229 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.036776 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-79mqm"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.045070 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.048944 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wjgj7"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.061245 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.061503 4751 scope.go:117] "RemoveContainer" containerID="2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.062851 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mfdl4"] Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.079917 4751 scope.go:117] "RemoveContainer" containerID="e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.080443 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d\": container with ID starting with e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d not found: ID does not exist" containerID="e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.080542 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d"} err="failed to get container status \"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d\": rpc error: code = NotFound desc = could not find container \"e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d\": container with ID starting with e42239a1d931d3fbfd5ffc224b0b2237ad320c0a542ac54f5627b78b48626f6d not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.080641 4751 scope.go:117] "RemoveContainer" containerID="e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.081362 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880\": container with ID starting with e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880 not found: ID does not exist" containerID="e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.081412 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880"} err="failed to get container status \"e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880\": rpc error: code = NotFound desc = could not find container \"e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880\": container with ID starting with e5c5e35e20e80196091559da44b6e2c6bca6bd4c3c1111a1054257cfddf36880 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.081435 4751 scope.go:117] "RemoveContainer" containerID="2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.081814 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9\": container with ID starting with 2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9 not found: ID does not exist" containerID="2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.081873 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9"} err="failed to get container status \"2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9\": rpc error: code = NotFound desc = could not find container \"2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9\": container with ID starting with 2ab3657cf5e8eb40f9f6e9452d920b74db6bf422fbe9a159573b1e0b1046acd9 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.081912 4751 scope.go:117] "RemoveContainer" containerID="fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.097299 4751 scope.go:117] "RemoveContainer" containerID="49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.112404 4751 scope.go:117] "RemoveContainer" containerID="59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.127153 4751 scope.go:117] "RemoveContainer" containerID="fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.127913 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14\": container with ID starting with fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14 not found: ID does not exist" containerID="fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.127957 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14"} err="failed to get container status \"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14\": rpc error: code = NotFound desc = could not find container \"fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14\": container with ID starting with fa58e7f4afb01f980a2869e22327291f5cd47cdd20c831646d3843178dfc4d14 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.127991 4751 scope.go:117] "RemoveContainer" containerID="49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.128643 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8\": container with ID starting with 49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8 not found: ID does not exist" containerID="49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.128674 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8"} err="failed to get container status \"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8\": rpc error: code = NotFound desc = could not find container \"49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8\": container with ID starting with 49e9654a1879e973d90520175c539cb3da5c75b184c018452a7fc1c1cba208e8 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.128694 4751 scope.go:117] "RemoveContainer" containerID="59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.129051 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569\": container with ID starting with 59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569 not found: ID does not exist" containerID="59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.129110 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569"} err="failed to get container status \"59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569\": rpc error: code = NotFound desc = could not find container \"59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569\": container with ID starting with 59d2c70ae32b25aa191f44319ce5ee377b5b8542a675880cabcdcdef923e1569 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.129213 4751 scope.go:117] "RemoveContainer" containerID="814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.144876 4751 scope.go:117] "RemoveContainer" containerID="42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.164731 4751 scope.go:117] "RemoveContainer" containerID="5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.186906 4751 scope.go:117] "RemoveContainer" containerID="814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.187452 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c\": container with ID starting with 814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c not found: ID does not exist" containerID="814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.187497 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c"} err="failed to get container status \"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c\": rpc error: code = NotFound desc = could not find container \"814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c\": container with ID starting with 814c6ce59fad8deb933a5709eb1fde917da886dd6e46967a29f1548024ea173c not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.187536 4751 scope.go:117] "RemoveContainer" containerID="42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.187894 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5\": container with ID starting with 42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5 not found: ID does not exist" containerID="42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.187914 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5"} err="failed to get container status \"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5\": rpc error: code = NotFound desc = could not find container \"42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5\": container with ID starting with 42b90eb9fdaafb5c64fa33abcef267f60b3c3440f73a70d6152293edeef2dae5 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.187930 4751 scope.go:117] "RemoveContainer" containerID="5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.188259 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0\": container with ID starting with 5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0 not found: ID does not exist" containerID="5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.188280 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0"} err="failed to get container status \"5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0\": rpc error: code = NotFound desc = could not find container \"5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0\": container with ID starting with 5abf21ffa185fc22a5b665f8b416e196dd07c7d1266424e326bdfe6f3b74d2e0 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.188294 4751 scope.go:117] "RemoveContainer" containerID="e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.203450 4751 scope.go:117] "RemoveContainer" containerID="3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.236860 4751 scope.go:117] "RemoveContainer" containerID="3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.251361 4751 scope.go:117] "RemoveContainer" containerID="e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.252023 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888\": container with ID starting with e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888 not found: ID does not exist" containerID="e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.252060 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888"} err="failed to get container status \"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888\": rpc error: code = NotFound desc = could not find container \"e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888\": container with ID starting with e316ab8041c92b6a8b2caf6fa9633dc72422c6932d506c35ad3f9573692a6888 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.252089 4751 scope.go:117] "RemoveContainer" containerID="3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.252816 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871\": container with ID starting with 3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871 not found: ID does not exist" containerID="3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.252895 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871"} err="failed to get container status \"3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871\": rpc error: code = NotFound desc = could not find container \"3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871\": container with ID starting with 3886ab7efa9661f6a584252b301433825151b7bb0141c2dcfebe82e981f83871 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.252951 4751 scope.go:117] "RemoveContainer" containerID="3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594" Oct 02 10:58:33 crc kubenswrapper[4751]: E1002 10:58:33.253395 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594\": container with ID starting with 3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594 not found: ID does not exist" containerID="3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.253497 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594"} err="failed to get container status \"3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594\": rpc error: code = NotFound desc = could not find container \"3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594\": container with ID starting with 3161138d1c05256196ee5181abcb7104041d092d8b28c1657a71abce8ee28594 not found: ID does not exist" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.561667 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" path="/var/lib/kubelet/pods/34f2464f-57f2-443e-8221-e21436dc9db0/volumes" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.562795 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" path="/var/lib/kubelet/pods/3adadf57-57a6-4200-9890-6b3abcf24663/volumes" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.563832 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" path="/var/lib/kubelet/pods/4257aeff-6e4a-4b1e-ae09-64065d3eaba8/volumes" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.565695 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" path="/var/lib/kubelet/pods/4718f397-506a-4690-8942-caecf3a3e6c6/volumes" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.566397 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" path="/var/lib/kubelet/pods/fbf7ed8c-d710-4d79-bdfc-31a84398d154/volumes" Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.995303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" event={"ID":"94dce789-2750-40a1-b622-11ad58a438df","Type":"ContainerStarted","Data":"effeae5c9fd2b54e7d649d36afe952db8634f434ee65953c628ff45718b6e395"} Oct 02 10:58:33 crc kubenswrapper[4751]: I1002 10:58:33.995675 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.001308 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.015406 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vdm72" podStartSLOduration=2.015388913 podStartE2EDuration="2.015388913s" podCreationTimestamp="2025-10-02 10:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:58:34.013047366 +0000 UTC m=+396.067273816" watchObservedRunningTime="2025-10-02 10:58:34.015388913 +0000 UTC m=+396.069615363" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291021 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bxscj"] Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291678 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291693 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291711 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291719 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291730 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291738 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291747 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291754 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291765 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291772 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291798 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291807 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291822 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291830 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291843 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291850 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="extract-utilities" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291859 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291867 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291875 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291882 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291890 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291898 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291905 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291915 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="extract-content" Oct 02 10:58:34 crc kubenswrapper[4751]: E1002 10:58:34.291929 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.291936 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292043 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4718f397-506a-4690-8942-caecf3a3e6c6" containerName="marketplace-operator" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292057 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4257aeff-6e4a-4b1e-ae09-64065d3eaba8" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292068 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adadf57-57a6-4200-9890-6b3abcf24663" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292079 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf7ed8c-d710-4d79-bdfc-31a84398d154" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292089 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f2464f-57f2-443e-8221-e21436dc9db0" containerName="registry-server" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.292943 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.294823 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.299570 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bxscj"] Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.431230 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-utilities\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.431302 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26d82\" (UniqueName: \"kubernetes.io/projected/4633ba02-9f3d-4de1-8214-0e908d7d3b72-kube-api-access-26d82\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.431437 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-catalog-content\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.475405 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cqd2n"] Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.476499 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.479832 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.487193 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqd2n"] Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.532870 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-catalog-content\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.532992 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-utilities\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533069 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26d82\" (UniqueName: \"kubernetes.io/projected/4633ba02-9f3d-4de1-8214-0e908d7d3b72-kube-api-access-26d82\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533205 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbw28\" (UniqueName: \"kubernetes.io/projected/9b28c6e5-2f16-4e4c-8491-f0c685d33817-kube-api-access-mbw28\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533246 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-utilities\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533309 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-catalog-content\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533552 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-utilities\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.533822 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4633ba02-9f3d-4de1-8214-0e908d7d3b72-catalog-content\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.550635 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26d82\" (UniqueName: \"kubernetes.io/projected/4633ba02-9f3d-4de1-8214-0e908d7d3b72-kube-api-access-26d82\") pod \"redhat-marketplace-bxscj\" (UID: \"4633ba02-9f3d-4de1-8214-0e908d7d3b72\") " pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.624586 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.634335 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbw28\" (UniqueName: \"kubernetes.io/projected/9b28c6e5-2f16-4e4c-8491-f0c685d33817-kube-api-access-mbw28\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.634397 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-utilities\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.634433 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-catalog-content\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.634869 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-utilities\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.634899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b28c6e5-2f16-4e4c-8491-f0c685d33817-catalog-content\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.653303 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbw28\" (UniqueName: \"kubernetes.io/projected/9b28c6e5-2f16-4e4c-8491-f0c685d33817-kube-api-access-mbw28\") pod \"redhat-operators-cqd2n\" (UID: \"9b28c6e5-2f16-4e4c-8491-f0c685d33817\") " pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.810235 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.835084 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bxscj"] Oct 02 10:58:34 crc kubenswrapper[4751]: W1002 10:58:34.842026 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4633ba02_9f3d_4de1_8214_0e908d7d3b72.slice/crio-82d8d522d1c0619b507b31fdb29576dde2df9a8db511f432eb41f5b5884d8adc WatchSource:0}: Error finding container 82d8d522d1c0619b507b31fdb29576dde2df9a8db511f432eb41f5b5884d8adc: Status 404 returned error can't find the container with id 82d8d522d1c0619b507b31fdb29576dde2df9a8db511f432eb41f5b5884d8adc Oct 02 10:58:34 crc kubenswrapper[4751]: I1002 10:58:34.987942 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cqd2n"] Oct 02 10:58:35 crc kubenswrapper[4751]: I1002 10:58:35.008036 4751 generic.go:334] "Generic (PLEG): container finished" podID="4633ba02-9f3d-4de1-8214-0e908d7d3b72" containerID="c441180aa0747a250fd483f15971bd59e32d866c41b1be06667d2b0cff252a3e" exitCode=0 Oct 02 10:58:35 crc kubenswrapper[4751]: I1002 10:58:35.008110 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bxscj" event={"ID":"4633ba02-9f3d-4de1-8214-0e908d7d3b72","Type":"ContainerDied","Data":"c441180aa0747a250fd483f15971bd59e32d866c41b1be06667d2b0cff252a3e"} Oct 02 10:58:35 crc kubenswrapper[4751]: I1002 10:58:35.008245 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bxscj" event={"ID":"4633ba02-9f3d-4de1-8214-0e908d7d3b72","Type":"ContainerStarted","Data":"82d8d522d1c0619b507b31fdb29576dde2df9a8db511f432eb41f5b5884d8adc"} Oct 02 10:58:35 crc kubenswrapper[4751]: W1002 10:58:35.065730 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b28c6e5_2f16_4e4c_8491_f0c685d33817.slice/crio-22d7149f70a3994703a12f152be8dcd6d3dd41c0806accd0e3436d2185e1fb7f WatchSource:0}: Error finding container 22d7149f70a3994703a12f152be8dcd6d3dd41c0806accd0e3436d2185e1fb7f: Status 404 returned error can't find the container with id 22d7149f70a3994703a12f152be8dcd6d3dd41c0806accd0e3436d2185e1fb7f Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.019514 4751 generic.go:334] "Generic (PLEG): container finished" podID="4633ba02-9f3d-4de1-8214-0e908d7d3b72" containerID="ae9438b9b45c8fb40d37eee3a00e8ee52b14613035ed9348e7d7debb9731bf58" exitCode=0 Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.019575 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bxscj" event={"ID":"4633ba02-9f3d-4de1-8214-0e908d7d3b72","Type":"ContainerDied","Data":"ae9438b9b45c8fb40d37eee3a00e8ee52b14613035ed9348e7d7debb9731bf58"} Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.022451 4751 generic.go:334] "Generic (PLEG): container finished" podID="9b28c6e5-2f16-4e4c-8491-f0c685d33817" containerID="2a535dac08a39604cdf20a188e6336d44f9e1636e8e4a415166a965fa69ce5d7" exitCode=0 Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.022535 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqd2n" event={"ID":"9b28c6e5-2f16-4e4c-8491-f0c685d33817","Type":"ContainerDied","Data":"2a535dac08a39604cdf20a188e6336d44f9e1636e8e4a415166a965fa69ce5d7"} Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.022602 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqd2n" event={"ID":"9b28c6e5-2f16-4e4c-8491-f0c685d33817","Type":"ContainerStarted","Data":"22d7149f70a3994703a12f152be8dcd6d3dd41c0806accd0e3436d2185e1fb7f"} Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.672478 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bhk57"] Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.674000 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.681046 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.681428 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhk57"] Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.858040 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-catalog-content\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.858299 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-utilities\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.858387 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knx7v\" (UniqueName: \"kubernetes.io/projected/f17f0af2-1997-416f-a908-4653c87513a8-kube-api-access-knx7v\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.877119 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-85txs"] Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.878019 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.881138 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.887980 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85txs"] Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.959217 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knx7v\" (UniqueName: \"kubernetes.io/projected/f17f0af2-1997-416f-a908-4653c87513a8-kube-api-access-knx7v\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.959301 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-catalog-content\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.959350 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-utilities\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.959809 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-utilities\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.960192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17f0af2-1997-416f-a908-4653c87513a8-catalog-content\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.979063 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knx7v\" (UniqueName: \"kubernetes.io/projected/f17f0af2-1997-416f-a908-4653c87513a8-kube-api-access-knx7v\") pod \"certified-operators-bhk57\" (UID: \"f17f0af2-1997-416f-a908-4653c87513a8\") " pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:36 crc kubenswrapper[4751]: I1002 10:58:36.989568 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.031083 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqd2n" event={"ID":"9b28c6e5-2f16-4e4c-8491-f0c685d33817","Type":"ContainerStarted","Data":"9b436c9126e32e16268a88cbcb32c5d0f4de22189d40ac537a7b2a31c195c026"} Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.037329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bxscj" event={"ID":"4633ba02-9f3d-4de1-8214-0e908d7d3b72","Type":"ContainerStarted","Data":"2183100bf56622586abda824b5fb3d662bb07ff558f4728791f45abba4d32b47"} Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.060162 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-utilities\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.060270 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z68jn\" (UniqueName: \"kubernetes.io/projected/3af1207d-dab1-4b06-9dc4-0cfed43f145f-kube-api-access-z68jn\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.060303 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-catalog-content\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.074935 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bxscj" podStartSLOduration=1.6554780230000001 podStartE2EDuration="3.074915393s" podCreationTimestamp="2025-10-02 10:58:34 +0000 UTC" firstStartedPulling="2025-10-02 10:58:35.009808859 +0000 UTC m=+397.064035309" lastFinishedPulling="2025-10-02 10:58:36.429246229 +0000 UTC m=+398.483472679" observedRunningTime="2025-10-02 10:58:37.072200406 +0000 UTC m=+399.126426856" watchObservedRunningTime="2025-10-02 10:58:37.074915393 +0000 UTC m=+399.129141843" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.162255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-catalog-content\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.162621 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-utilities\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.162675 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z68jn\" (UniqueName: \"kubernetes.io/projected/3af1207d-dab1-4b06-9dc4-0cfed43f145f-kube-api-access-z68jn\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.164360 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-catalog-content\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.164663 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af1207d-dab1-4b06-9dc4-0cfed43f145f-utilities\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.179752 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z68jn\" (UniqueName: \"kubernetes.io/projected/3af1207d-dab1-4b06-9dc4-0cfed43f145f-kube-api-access-z68jn\") pod \"community-operators-85txs\" (UID: \"3af1207d-dab1-4b06-9dc4-0cfed43f145f\") " pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.221102 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.393931 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhk57"] Oct 02 10:58:37 crc kubenswrapper[4751]: I1002 10:58:37.420570 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-85txs"] Oct 02 10:58:37 crc kubenswrapper[4751]: W1002 10:58:37.426920 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3af1207d_dab1_4b06_9dc4_0cfed43f145f.slice/crio-d04ae2fc14ff36eb2659dfd761c5d8ef66f2bdc5779f88de263c69839a5f3424 WatchSource:0}: Error finding container d04ae2fc14ff36eb2659dfd761c5d8ef66f2bdc5779f88de263c69839a5f3424: Status 404 returned error can't find the container with id d04ae2fc14ff36eb2659dfd761c5d8ef66f2bdc5779f88de263c69839a5f3424 Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.045459 4751 generic.go:334] "Generic (PLEG): container finished" podID="9b28c6e5-2f16-4e4c-8491-f0c685d33817" containerID="9b436c9126e32e16268a88cbcb32c5d0f4de22189d40ac537a7b2a31c195c026" exitCode=0 Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.045562 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqd2n" event={"ID":"9b28c6e5-2f16-4e4c-8491-f0c685d33817","Type":"ContainerDied","Data":"9b436c9126e32e16268a88cbcb32c5d0f4de22189d40ac537a7b2a31c195c026"} Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.048737 4751 generic.go:334] "Generic (PLEG): container finished" podID="f17f0af2-1997-416f-a908-4653c87513a8" containerID="2015b0948dff8e0bd27f223f09070c23c52f98a41d66544cf23cf9117116a44d" exitCode=0 Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.048878 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhk57" event={"ID":"f17f0af2-1997-416f-a908-4653c87513a8","Type":"ContainerDied","Data":"2015b0948dff8e0bd27f223f09070c23c52f98a41d66544cf23cf9117116a44d"} Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.048911 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhk57" event={"ID":"f17f0af2-1997-416f-a908-4653c87513a8","Type":"ContainerStarted","Data":"f9a403437cf6c660e1b7e2dc70dc62bad1b4fa0fa01301086bf8455fc7f68ea6"} Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.052130 4751 generic.go:334] "Generic (PLEG): container finished" podID="3af1207d-dab1-4b06-9dc4-0cfed43f145f" containerID="48ed9c6fe6b543eced7591ee889e5838420d5f84daac96ed58c6ae7d2c5a4ba7" exitCode=0 Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.052213 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85txs" event={"ID":"3af1207d-dab1-4b06-9dc4-0cfed43f145f","Type":"ContainerDied","Data":"48ed9c6fe6b543eced7591ee889e5838420d5f84daac96ed58c6ae7d2c5a4ba7"} Oct 02 10:58:38 crc kubenswrapper[4751]: I1002 10:58:38.052343 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85txs" event={"ID":"3af1207d-dab1-4b06-9dc4-0cfed43f145f","Type":"ContainerStarted","Data":"d04ae2fc14ff36eb2659dfd761c5d8ef66f2bdc5779f88de263c69839a5f3424"} Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.062522 4751 generic.go:334] "Generic (PLEG): container finished" podID="3af1207d-dab1-4b06-9dc4-0cfed43f145f" containerID="5ddc8ceee11b235d204635e635f3d40ea386ed9390877d3901a0bd01afd1c14e" exitCode=0 Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.062568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85txs" event={"ID":"3af1207d-dab1-4b06-9dc4-0cfed43f145f","Type":"ContainerDied","Data":"5ddc8ceee11b235d204635e635f3d40ea386ed9390877d3901a0bd01afd1c14e"} Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.065385 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cqd2n" event={"ID":"9b28c6e5-2f16-4e4c-8491-f0c685d33817","Type":"ContainerStarted","Data":"a2b295705a77ead50fe471b8c342010f8441b56fe3b2666209e91650a8f77859"} Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.068711 4751 generic.go:334] "Generic (PLEG): container finished" podID="f17f0af2-1997-416f-a908-4653c87513a8" containerID="2d98913722597c4a09184fc7c0d4cf247293560517f2291bac69f369a8a415a2" exitCode=0 Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.068757 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhk57" event={"ID":"f17f0af2-1997-416f-a908-4653c87513a8","Type":"ContainerDied","Data":"2d98913722597c4a09184fc7c0d4cf247293560517f2291bac69f369a8a415a2"} Oct 02 10:58:40 crc kubenswrapper[4751]: I1002 10:58:40.103435 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cqd2n" podStartSLOduration=3.62323461 podStartE2EDuration="6.103410614s" podCreationTimestamp="2025-10-02 10:58:34 +0000 UTC" firstStartedPulling="2025-10-02 10:58:36.02413274 +0000 UTC m=+398.078359230" lastFinishedPulling="2025-10-02 10:58:38.504308784 +0000 UTC m=+400.558535234" observedRunningTime="2025-10-02 10:58:40.100986986 +0000 UTC m=+402.155213436" watchObservedRunningTime="2025-10-02 10:58:40.103410614 +0000 UTC m=+402.157637104" Oct 02 10:58:42 crc kubenswrapper[4751]: I1002 10:58:42.079470 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-85txs" event={"ID":"3af1207d-dab1-4b06-9dc4-0cfed43f145f","Type":"ContainerStarted","Data":"01d0a5b863f95348aa5ab2e91d3c7b2d93139d07f77d96045291dd222de31537"} Oct 02 10:58:42 crc kubenswrapper[4751]: I1002 10:58:42.082059 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhk57" event={"ID":"f17f0af2-1997-416f-a908-4653c87513a8","Type":"ContainerStarted","Data":"d86b2b85acbaca19c534acc08e2294a471c40e2cb4b9df671f5598092aaa0b66"} Oct 02 10:58:42 crc kubenswrapper[4751]: I1002 10:58:42.094311 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-85txs" podStartSLOduration=2.652569046 podStartE2EDuration="6.094296715s" podCreationTimestamp="2025-10-02 10:58:36 +0000 UTC" firstStartedPulling="2025-10-02 10:58:38.053748118 +0000 UTC m=+400.107974568" lastFinishedPulling="2025-10-02 10:58:41.495475787 +0000 UTC m=+403.549702237" observedRunningTime="2025-10-02 10:58:42.093296776 +0000 UTC m=+404.147523226" watchObservedRunningTime="2025-10-02 10:58:42.094296715 +0000 UTC m=+404.148523155" Oct 02 10:58:42 crc kubenswrapper[4751]: I1002 10:58:42.115113 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bhk57" podStartSLOduration=3.323407523 podStartE2EDuration="6.115095694s" podCreationTimestamp="2025-10-02 10:58:36 +0000 UTC" firstStartedPulling="2025-10-02 10:58:38.051622867 +0000 UTC m=+400.105849317" lastFinishedPulling="2025-10-02 10:58:40.843311038 +0000 UTC m=+402.897537488" observedRunningTime="2025-10-02 10:58:42.113917411 +0000 UTC m=+404.168143861" watchObservedRunningTime="2025-10-02 10:58:42.115095694 +0000 UTC m=+404.169322134" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.625211 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.625539 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.661265 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.810672 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.810918 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:44 crc kubenswrapper[4751]: I1002 10:58:44.875344 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:45 crc kubenswrapper[4751]: I1002 10:58:45.155485 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cqd2n" Oct 02 10:58:45 crc kubenswrapper[4751]: I1002 10:58:45.159730 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bxscj" Oct 02 10:58:46 crc kubenswrapper[4751]: I1002 10:58:46.990724 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:46 crc kubenswrapper[4751]: I1002 10:58:46.991101 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:47 crc kubenswrapper[4751]: I1002 10:58:47.027225 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:47 crc kubenswrapper[4751]: I1002 10:58:47.147747 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bhk57" Oct 02 10:58:47 crc kubenswrapper[4751]: I1002 10:58:47.222161 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:47 crc kubenswrapper[4751]: I1002 10:58:47.222214 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:47 crc kubenswrapper[4751]: I1002 10:58:47.255961 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:48 crc kubenswrapper[4751]: I1002 10:58:48.187562 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-85txs" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.566286 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mj9xp"] Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.568062 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.588129 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mj9xp"] Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748302 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d16ee08f-81fa-40a5-b735-b85e9146718e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748375 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-trusted-ca\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748440 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g77jz\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-kube-api-access-g77jz\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d16ee08f-81fa-40a5-b735-b85e9146718e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748616 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748670 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-certificates\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748706 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-tls\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.748781 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-bound-sa-token\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.780306 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850734 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d16ee08f-81fa-40a5-b735-b85e9146718e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850802 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-trusted-ca\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850857 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g77jz\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-kube-api-access-g77jz\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850901 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d16ee08f-81fa-40a5-b735-b85e9146718e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850951 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-certificates\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.850994 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-tls\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.851075 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-bound-sa-token\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.851218 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d16ee08f-81fa-40a5-b735-b85e9146718e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.852366 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-certificates\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.853137 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d16ee08f-81fa-40a5-b735-b85e9146718e-trusted-ca\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.859105 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-registry-tls\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.859221 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d16ee08f-81fa-40a5-b735-b85e9146718e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.867221 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g77jz\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-kube-api-access-g77jz\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.867316 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d16ee08f-81fa-40a5-b735-b85e9146718e-bound-sa-token\") pod \"image-registry-66df7c8f76-mj9xp\" (UID: \"d16ee08f-81fa-40a5-b735-b85e9146718e\") " pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:58 crc kubenswrapper[4751]: I1002 10:58:58.883034 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:58:59 crc kubenswrapper[4751]: I1002 10:58:59.067289 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mj9xp"] Oct 02 10:58:59 crc kubenswrapper[4751]: W1002 10:58:59.071000 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd16ee08f_81fa_40a5_b735_b85e9146718e.slice/crio-6054054253f9f5bc7e6228051067b9cfefd422b2abd2f5ee5ad2f273295209e0 WatchSource:0}: Error finding container 6054054253f9f5bc7e6228051067b9cfefd422b2abd2f5ee5ad2f273295209e0: Status 404 returned error can't find the container with id 6054054253f9f5bc7e6228051067b9cfefd422b2abd2f5ee5ad2f273295209e0 Oct 02 10:58:59 crc kubenswrapper[4751]: I1002 10:58:59.180006 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" event={"ID":"d16ee08f-81fa-40a5-b735-b85e9146718e","Type":"ContainerStarted","Data":"6054054253f9f5bc7e6228051067b9cfefd422b2abd2f5ee5ad2f273295209e0"} Oct 02 10:59:00 crc kubenswrapper[4751]: I1002 10:59:00.189599 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" event={"ID":"d16ee08f-81fa-40a5-b735-b85e9146718e","Type":"ContainerStarted","Data":"1fc0a9e5b90f5ece79014b422840253f0b4d18863e8e15bdeb549ec3196fbb98"} Oct 02 10:59:00 crc kubenswrapper[4751]: I1002 10:59:00.190138 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:59:00 crc kubenswrapper[4751]: I1002 10:59:00.222380 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" podStartSLOduration=2.222356802 podStartE2EDuration="2.222356802s" podCreationTimestamp="2025-10-02 10:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:59:00.219226413 +0000 UTC m=+422.273452903" watchObservedRunningTime="2025-10-02 10:59:00.222356802 +0000 UTC m=+422.276583292" Oct 02 10:59:01 crc kubenswrapper[4751]: I1002 10:59:01.507376 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:59:01 crc kubenswrapper[4751]: I1002 10:59:01.507457 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:59:01 crc kubenswrapper[4751]: I1002 10:59:01.507518 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 10:59:01 crc kubenswrapper[4751]: I1002 10:59:01.508320 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:59:01 crc kubenswrapper[4751]: I1002 10:59:01.508428 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49" gracePeriod=600 Oct 02 10:59:02 crc kubenswrapper[4751]: I1002 10:59:02.205342 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49" exitCode=0 Oct 02 10:59:02 crc kubenswrapper[4751]: I1002 10:59:02.205407 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49"} Oct 02 10:59:02 crc kubenswrapper[4751]: I1002 10:59:02.206027 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a"} Oct 02 10:59:02 crc kubenswrapper[4751]: I1002 10:59:02.206059 4751 scope.go:117] "RemoveContainer" containerID="c41e86565c4686ac39b843f1b611d53e14bdcfd3e0967f3bc8b87056d6b0e1c0" Oct 02 10:59:18 crc kubenswrapper[4751]: I1002 10:59:18.890273 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mj9xp" Oct 02 10:59:18 crc kubenswrapper[4751]: I1002 10:59:18.945824 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:59:43 crc kubenswrapper[4751]: I1002 10:59:43.994264 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" podUID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" containerName="registry" containerID="cri-o://2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe" gracePeriod=30 Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.367991 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.498095 4751 generic.go:334] "Generic (PLEG): container finished" podID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" containerID="2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe" exitCode=0 Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.498141 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" event={"ID":"f27e1b4b-501e-4721-ba3e-c3b3d862e53c","Type":"ContainerDied","Data":"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe"} Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.498206 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" event={"ID":"f27e1b4b-501e-4721-ba3e-c3b3d862e53c","Type":"ContainerDied","Data":"03875ba66e8edc9604ea3ce614bfa91c437a303f58a13914beca02bd5c79d647"} Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.498225 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qg59s" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.498228 4751 scope.go:117] "RemoveContainer" containerID="2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.518492 4751 scope.go:117] "RemoveContainer" containerID="2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe" Oct 02 10:59:44 crc kubenswrapper[4751]: E1002 10:59:44.518910 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe\": container with ID starting with 2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe not found: ID does not exist" containerID="2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.518969 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe"} err="failed to get container status \"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe\": rpc error: code = NotFound desc = could not find container \"2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe\": container with ID starting with 2e99416d48eeec5c0d04a8f655781eb89e681d7769d3d8ac79e1a39c25a0a0fe not found: ID does not exist" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532326 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532390 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532443 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mmjw\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532517 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532572 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532607 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532776 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.532840 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca\") pod \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\" (UID: \"f27e1b4b-501e-4721-ba3e-c3b3d862e53c\") " Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.534631 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.534656 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.540992 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw" (OuterVolumeSpecName: "kube-api-access-2mmjw") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "kube-api-access-2mmjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.542067 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.543116 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.543967 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.554492 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.572857 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f27e1b4b-501e-4721-ba3e-c3b3d862e53c" (UID: "f27e1b4b-501e-4721-ba3e-c3b3d862e53c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.634773 4751 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.635536 4751 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.635757 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mmjw\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-kube-api-access-2mmjw\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.635967 4751 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.636165 4751 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.636697 4751 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.636856 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f27e1b4b-501e-4721-ba3e-c3b3d862e53c-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.850204 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:59:44 crc kubenswrapper[4751]: I1002 10:59:44.855145 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qg59s"] Oct 02 10:59:45 crc kubenswrapper[4751]: I1002 10:59:45.563246 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" path="/var/lib/kubelet/pods/f27e1b4b-501e-4721-ba3e-c3b3d862e53c/volumes" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.148724 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t"] Oct 02 11:00:00 crc kubenswrapper[4751]: E1002 11:00:00.149655 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" containerName="registry" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.149676 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" containerName="registry" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.149855 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27e1b4b-501e-4721-ba3e-c3b3d862e53c" containerName="registry" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.150520 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.153063 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.153434 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.158448 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t"] Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.240068 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rzr6\" (UniqueName: \"kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.240130 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.240189 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.340755 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rzr6\" (UniqueName: \"kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.340852 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.340890 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.341863 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.349630 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.357641 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rzr6\" (UniqueName: \"kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6\") pod \"collect-profiles-29323380-kxc4t\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.473313 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:00 crc kubenswrapper[4751]: I1002 11:00:00.734017 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t"] Oct 02 11:00:01 crc kubenswrapper[4751]: I1002 11:00:01.617505 4751 generic.go:334] "Generic (PLEG): container finished" podID="c25d4150-62c8-4ec1-ae32-050a04eafd3e" containerID="3dea4feb9fb491e1a03c581a7fa1f0ed4c011fb19ed0a232bcce1c2642878209" exitCode=0 Oct 02 11:00:01 crc kubenswrapper[4751]: I1002 11:00:01.617576 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" event={"ID":"c25d4150-62c8-4ec1-ae32-050a04eafd3e","Type":"ContainerDied","Data":"3dea4feb9fb491e1a03c581a7fa1f0ed4c011fb19ed0a232bcce1c2642878209"} Oct 02 11:00:01 crc kubenswrapper[4751]: I1002 11:00:01.617789 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" event={"ID":"c25d4150-62c8-4ec1-ae32-050a04eafd3e","Type":"ContainerStarted","Data":"4099ae08a3e41a36cb2c8e08eb05320a5668648a13d64859f0c1cc323e77c725"} Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.870902 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.983448 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume\") pod \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.983529 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume\") pod \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.983646 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rzr6\" (UniqueName: \"kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6\") pod \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\" (UID: \"c25d4150-62c8-4ec1-ae32-050a04eafd3e\") " Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.985065 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c25d4150-62c8-4ec1-ae32-050a04eafd3e" (UID: "c25d4150-62c8-4ec1-ae32-050a04eafd3e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.988739 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6" (OuterVolumeSpecName: "kube-api-access-7rzr6") pod "c25d4150-62c8-4ec1-ae32-050a04eafd3e" (UID: "c25d4150-62c8-4ec1-ae32-050a04eafd3e"). InnerVolumeSpecName "kube-api-access-7rzr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:00:02 crc kubenswrapper[4751]: I1002 11:00:02.988894 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c25d4150-62c8-4ec1-ae32-050a04eafd3e" (UID: "c25d4150-62c8-4ec1-ae32-050a04eafd3e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.084860 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c25d4150-62c8-4ec1-ae32-050a04eafd3e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.084893 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rzr6\" (UniqueName: \"kubernetes.io/projected/c25d4150-62c8-4ec1-ae32-050a04eafd3e-kube-api-access-7rzr6\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.084905 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c25d4150-62c8-4ec1-ae32-050a04eafd3e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.632384 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" event={"ID":"c25d4150-62c8-4ec1-ae32-050a04eafd3e","Type":"ContainerDied","Data":"4099ae08a3e41a36cb2c8e08eb05320a5668648a13d64859f0c1cc323e77c725"} Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.632431 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4099ae08a3e41a36cb2c8e08eb05320a5668648a13d64859f0c1cc323e77c725" Oct 02 11:00:03 crc kubenswrapper[4751]: I1002 11:00:03.632486 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t" Oct 02 11:01:01 crc kubenswrapper[4751]: I1002 11:01:01.506962 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:01:01 crc kubenswrapper[4751]: I1002 11:01:01.507699 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:01:31 crc kubenswrapper[4751]: I1002 11:01:31.506813 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:01:31 crc kubenswrapper[4751]: I1002 11:01:31.507434 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:02:01 crc kubenswrapper[4751]: I1002 11:02:01.506750 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:02:01 crc kubenswrapper[4751]: I1002 11:02:01.507340 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:02:01 crc kubenswrapper[4751]: I1002 11:02:01.507391 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:02:01 crc kubenswrapper[4751]: I1002 11:02:01.507917 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:02:01 crc kubenswrapper[4751]: I1002 11:02:01.507972 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a" gracePeriod=600 Oct 02 11:02:02 crc kubenswrapper[4751]: I1002 11:02:02.406305 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a" exitCode=0 Oct 02 11:02:02 crc kubenswrapper[4751]: I1002 11:02:02.406366 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a"} Oct 02 11:02:02 crc kubenswrapper[4751]: I1002 11:02:02.406743 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed"} Oct 02 11:02:02 crc kubenswrapper[4751]: I1002 11:02:02.406778 4751 scope.go:117] "RemoveContainer" containerID="41e0aa075132e808db8de62a81ad3515e38c9d16c7792422718af891b1330f49" Oct 02 11:04:01 crc kubenswrapper[4751]: I1002 11:04:01.507124 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:04:01 crc kubenswrapper[4751]: I1002 11:04:01.507830 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:04:31 crc kubenswrapper[4751]: I1002 11:04:31.507866 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:04:31 crc kubenswrapper[4751]: I1002 11:04:31.508760 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:04:42 crc kubenswrapper[4751]: I1002 11:04:42.867117 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 11:04:42 crc kubenswrapper[4751]: I1002 11:04:42.868264 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerName="controller-manager" containerID="cri-o://4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb" gracePeriod=30 Oct 02 11:04:42 crc kubenswrapper[4751]: I1002 11:04:42.991929 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 11:04:42 crc kubenswrapper[4751]: I1002 11:04:42.992677 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerName="route-controller-manager" containerID="cri-o://18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c" gracePeriod=30 Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.221406 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.305194 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.390562 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca\") pod \"acd9ab77-c505-4404-95df-a0050d7e8a00\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.390608 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert\") pod \"acd9ab77-c505-4404-95df-a0050d7e8a00\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.390655 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config\") pod \"acd9ab77-c505-4404-95df-a0050d7e8a00\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.390680 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndbdq\" (UniqueName: \"kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq\") pod \"acd9ab77-c505-4404-95df-a0050d7e8a00\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.390731 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles\") pod \"acd9ab77-c505-4404-95df-a0050d7e8a00\" (UID: \"acd9ab77-c505-4404-95df-a0050d7e8a00\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.391563 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "acd9ab77-c505-4404-95df-a0050d7e8a00" (UID: "acd9ab77-c505-4404-95df-a0050d7e8a00"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.391645 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca" (OuterVolumeSpecName: "client-ca") pod "acd9ab77-c505-4404-95df-a0050d7e8a00" (UID: "acd9ab77-c505-4404-95df-a0050d7e8a00"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.392276 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config" (OuterVolumeSpecName: "config") pod "acd9ab77-c505-4404-95df-a0050d7e8a00" (UID: "acd9ab77-c505-4404-95df-a0050d7e8a00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.397869 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq" (OuterVolumeSpecName: "kube-api-access-ndbdq") pod "acd9ab77-c505-4404-95df-a0050d7e8a00" (UID: "acd9ab77-c505-4404-95df-a0050d7e8a00"). InnerVolumeSpecName "kube-api-access-ndbdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.398040 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "acd9ab77-c505-4404-95df-a0050d7e8a00" (UID: "acd9ab77-c505-4404-95df-a0050d7e8a00"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.412380 4751 generic.go:334] "Generic (PLEG): container finished" podID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerID="4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb" exitCode=0 Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.412431 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.412473 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" event={"ID":"acd9ab77-c505-4404-95df-a0050d7e8a00","Type":"ContainerDied","Data":"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb"} Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.412521 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-flnxn" event={"ID":"acd9ab77-c505-4404-95df-a0050d7e8a00","Type":"ContainerDied","Data":"01b4ac69b0def616b8c767b3f70136a9cafd1283637e3cd8f376709e7341b839"} Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.412542 4751 scope.go:117] "RemoveContainer" containerID="4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.415067 4751 generic.go:334] "Generic (PLEG): container finished" podID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerID="18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c" exitCode=0 Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.415116 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" event={"ID":"2d419bc9-8000-4cdc-bc74-be1147be91e9","Type":"ContainerDied","Data":"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c"} Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.415155 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.415735 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5" event={"ID":"2d419bc9-8000-4cdc-bc74-be1147be91e9","Type":"ContainerDied","Data":"d0127c188b8dd55e2ac09abc19b0dad3ff55089c126e50565c9601461e895d12"} Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.431205 4751 scope.go:117] "RemoveContainer" containerID="4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb" Oct 02 11:04:43 crc kubenswrapper[4751]: E1002 11:04:43.431926 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb\": container with ID starting with 4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb not found: ID does not exist" containerID="4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.432003 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb"} err="failed to get container status \"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb\": rpc error: code = NotFound desc = could not find container \"4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb\": container with ID starting with 4c848fb8cbef28d99dee332bf6a85f78ad1d1567f6e5916936b02fb52c42d4cb not found: ID does not exist" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.432056 4751 scope.go:117] "RemoveContainer" containerID="18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.446727 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.448268 4751 scope.go:117] "RemoveContainer" containerID="18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c" Oct 02 11:04:43 crc kubenswrapper[4751]: E1002 11:04:43.448889 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c\": container with ID starting with 18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c not found: ID does not exist" containerID="18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.448940 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c"} err="failed to get container status \"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c\": rpc error: code = NotFound desc = could not find container \"18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c\": container with ID starting with 18c2dfc46de3b6846e2e5cdbeabb98220398fff831fe599bcd8b8abb6ff5667c not found: ID does not exist" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.451215 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-flnxn"] Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492422 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert\") pod \"2d419bc9-8000-4cdc-bc74-be1147be91e9\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492495 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsw2b\" (UniqueName: \"kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b\") pod \"2d419bc9-8000-4cdc-bc74-be1147be91e9\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492531 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca\") pod \"2d419bc9-8000-4cdc-bc74-be1147be91e9\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492571 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config\") pod \"2d419bc9-8000-4cdc-bc74-be1147be91e9\" (UID: \"2d419bc9-8000-4cdc-bc74-be1147be91e9\") " Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492897 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492928 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acd9ab77-c505-4404-95df-a0050d7e8a00-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492944 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492959 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndbdq\" (UniqueName: \"kubernetes.io/projected/acd9ab77-c505-4404-95df-a0050d7e8a00-kube-api-access-ndbdq\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.492975 4751 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acd9ab77-c505-4404-95df-a0050d7e8a00-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.493548 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d419bc9-8000-4cdc-bc74-be1147be91e9" (UID: "2d419bc9-8000-4cdc-bc74-be1147be91e9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.493556 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config" (OuterVolumeSpecName: "config") pod "2d419bc9-8000-4cdc-bc74-be1147be91e9" (UID: "2d419bc9-8000-4cdc-bc74-be1147be91e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.496540 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d419bc9-8000-4cdc-bc74-be1147be91e9" (UID: "2d419bc9-8000-4cdc-bc74-be1147be91e9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.496999 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b" (OuterVolumeSpecName: "kube-api-access-wsw2b") pod "2d419bc9-8000-4cdc-bc74-be1147be91e9" (UID: "2d419bc9-8000-4cdc-bc74-be1147be91e9"). InnerVolumeSpecName "kube-api-access-wsw2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.558706 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" path="/var/lib/kubelet/pods/acd9ab77-c505-4404-95df-a0050d7e8a00/volumes" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.594091 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d419bc9-8000-4cdc-bc74-be1147be91e9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.594145 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsw2b\" (UniqueName: \"kubernetes.io/projected/2d419bc9-8000-4cdc-bc74-be1147be91e9-kube-api-access-wsw2b\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.594224 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.594254 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d419bc9-8000-4cdc-bc74-be1147be91e9-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.738156 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 11:04:43 crc kubenswrapper[4751]: I1002 11:04:43.744439 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fppx5"] Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.446393 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:44 crc kubenswrapper[4751]: E1002 11:04:44.446748 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25d4150-62c8-4ec1-ae32-050a04eafd3e" containerName="collect-profiles" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.446770 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25d4150-62c8-4ec1-ae32-050a04eafd3e" containerName="collect-profiles" Oct 02 11:04:44 crc kubenswrapper[4751]: E1002 11:04:44.446805 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerName="route-controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.446818 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerName="route-controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: E1002 11:04:44.446837 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerName="controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.446852 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerName="controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.447007 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd9ab77-c505-4404-95df-a0050d7e8a00" containerName="controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.447023 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" containerName="route-controller-manager" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.447053 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c25d4150-62c8-4ec1-ae32-050a04eafd3e" containerName="collect-profiles" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.447745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.450794 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.451410 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.451504 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5767455d47-22dsn"] Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.451733 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.452362 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.452566 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.452786 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.453144 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.456520 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.457019 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.457648 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.457698 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.457703 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.457876 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.467444 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.472307 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.475146 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5767455d47-22dsn"] Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605474 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/512063f5-5ac5-44a4-a66e-3d5562e21d6a-serving-cert\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605526 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605559 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qn6k\" (UniqueName: \"kubernetes.io/projected/512063f5-5ac5-44a4-a66e-3d5562e21d6a-kube-api-access-6qn6k\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605597 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605683 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605736 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-config\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605782 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-client-ca\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605814 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlgsl\" (UniqueName: \"kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.605846 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-proxy-ca-bundles\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706393 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/512063f5-5ac5-44a4-a66e-3d5562e21d6a-serving-cert\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706460 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qn6k\" (UniqueName: \"kubernetes.io/projected/512063f5-5ac5-44a4-a66e-3d5562e21d6a-kube-api-access-6qn6k\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706477 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706492 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706517 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-config\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-client-ca\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706563 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlgsl\" (UniqueName: \"kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.706600 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-proxy-ca-bundles\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.707621 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.707743 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-proxy-ca-bundles\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.707920 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-config\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.708000 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/512063f5-5ac5-44a4-a66e-3d5562e21d6a-client-ca\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.708029 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.711247 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.711325 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/512063f5-5ac5-44a4-a66e-3d5562e21d6a-serving-cert\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.723506 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlgsl\" (UniqueName: \"kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl\") pod \"route-controller-manager-7b5b9c6757-jpdnk\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.730003 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qn6k\" (UniqueName: \"kubernetes.io/projected/512063f5-5ac5-44a4-a66e-3d5562e21d6a-kube-api-access-6qn6k\") pod \"controller-manager-5767455d47-22dsn\" (UID: \"512063f5-5ac5-44a4-a66e-3d5562e21d6a\") " pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.772116 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.781397 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.821718 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:44 crc kubenswrapper[4751]: I1002 11:04:44.978963 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5767455d47-22dsn"] Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.038108 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:45 crc kubenswrapper[4751]: W1002 11:04:45.052358 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd417eb5f_e9ef_4ea8_aec0_ba7186157c69.slice/crio-2000e9a85a5ae063d1d113789d5b5d82b5d22569cf5217b5a4cae553a70f98e8 WatchSource:0}: Error finding container 2000e9a85a5ae063d1d113789d5b5d82b5d22569cf5217b5a4cae553a70f98e8: Status 404 returned error can't find the container with id 2000e9a85a5ae063d1d113789d5b5d82b5d22569cf5217b5a4cae553a70f98e8 Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.430835 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" event={"ID":"d417eb5f-e9ef-4ea8-aec0-ba7186157c69","Type":"ContainerStarted","Data":"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f"} Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.430897 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" event={"ID":"d417eb5f-e9ef-4ea8-aec0-ba7186157c69","Type":"ContainerStarted","Data":"2000e9a85a5ae063d1d113789d5b5d82b5d22569cf5217b5a4cae553a70f98e8"} Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.430980 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerName="route-controller-manager" containerID="cri-o://a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f" gracePeriod=30 Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.432308 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.432343 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" event={"ID":"512063f5-5ac5-44a4-a66e-3d5562e21d6a","Type":"ContainerStarted","Data":"f403d99bd67ae43787b03d2b213ee0c683c7d21bfff32ba48521dcecf45b989b"} Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.432360 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" event={"ID":"512063f5-5ac5-44a4-a66e-3d5562e21d6a","Type":"ContainerStarted","Data":"76e66a680c158c4274024044f03f9ce83f21928e8a9e0d28ef1d1799949e49a2"} Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.432557 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.437663 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.450810 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" podStartSLOduration=2.45079425 podStartE2EDuration="2.45079425s" podCreationTimestamp="2025-10-02 11:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:04:45.448462749 +0000 UTC m=+767.502689209" watchObservedRunningTime="2025-10-02 11:04:45.45079425 +0000 UTC m=+767.505020700" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.473138 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5767455d47-22dsn" podStartSLOduration=2.473123492 podStartE2EDuration="2.473123492s" podCreationTimestamp="2025-10-02 11:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:04:45.470330208 +0000 UTC m=+767.524556658" watchObservedRunningTime="2025-10-02 11:04:45.473123492 +0000 UTC m=+767.527349942" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.490370 4751 patch_prober.go:28] interesting pod/route-controller-manager-7b5b9c6757-jpdnk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": read tcp 10.217.0.2:57572->10.217.0.63:8443: read: connection reset by peer" start-of-body= Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.490426 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": read tcp 10.217.0.2:57572->10.217.0.63:8443: read: connection reset by peer" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.555700 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d419bc9-8000-4cdc-bc74-be1147be91e9" path="/var/lib/kubelet/pods/2d419bc9-8000-4cdc-bc74-be1147be91e9/volumes" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.723361 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-7b5b9c6757-jpdnk_d417eb5f-e9ef-4ea8-aec0-ba7186157c69/route-controller-manager/0.log" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.723653 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.924777 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config\") pod \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.924845 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlgsl\" (UniqueName: \"kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl\") pod \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.924904 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca\") pod \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.924929 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert\") pod \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\" (UID: \"d417eb5f-e9ef-4ea8-aec0-ba7186157c69\") " Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.926049 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca" (OuterVolumeSpecName: "client-ca") pod "d417eb5f-e9ef-4ea8-aec0-ba7186157c69" (UID: "d417eb5f-e9ef-4ea8-aec0-ba7186157c69"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.926271 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config" (OuterVolumeSpecName: "config") pod "d417eb5f-e9ef-4ea8-aec0-ba7186157c69" (UID: "d417eb5f-e9ef-4ea8-aec0-ba7186157c69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.931005 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl" (OuterVolumeSpecName: "kube-api-access-tlgsl") pod "d417eb5f-e9ef-4ea8-aec0-ba7186157c69" (UID: "d417eb5f-e9ef-4ea8-aec0-ba7186157c69"). InnerVolumeSpecName "kube-api-access-tlgsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:04:45 crc kubenswrapper[4751]: I1002 11:04:45.930990 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d417eb5f-e9ef-4ea8-aec0-ba7186157c69" (UID: "d417eb5f-e9ef-4ea8-aec0-ba7186157c69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.026410 4751 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.026444 4751 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.026453 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.026486 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlgsl\" (UniqueName: \"kubernetes.io/projected/d417eb5f-e9ef-4ea8-aec0-ba7186157c69-kube-api-access-tlgsl\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443258 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-7b5b9c6757-jpdnk_d417eb5f-e9ef-4ea8-aec0-ba7186157c69/route-controller-manager/0.log" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443315 4751 generic.go:334] "Generic (PLEG): container finished" podID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerID="a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f" exitCode=255 Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443638 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" event={"ID":"d417eb5f-e9ef-4ea8-aec0-ba7186157c69","Type":"ContainerDied","Data":"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f"} Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443822 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk" event={"ID":"d417eb5f-e9ef-4ea8-aec0-ba7186157c69","Type":"ContainerDied","Data":"2000e9a85a5ae063d1d113789d5b5d82b5d22569cf5217b5a4cae553a70f98e8"} Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.443846 4751 scope.go:117] "RemoveContainer" containerID="a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.449111 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr"] Oct 02 11:04:46 crc kubenswrapper[4751]: E1002 11:04:46.449537 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerName="route-controller-manager" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.449574 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerName="route-controller-manager" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.449837 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" containerName="route-controller-manager" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.451382 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.458020 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.458294 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.458422 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.458719 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.458899 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.459043 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.460710 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr"] Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.464506 4751 scope.go:117] "RemoveContainer" containerID="a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f" Oct 02 11:04:46 crc kubenswrapper[4751]: E1002 11:04:46.464928 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f\": container with ID starting with a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f not found: ID does not exist" containerID="a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.464953 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f"} err="failed to get container status \"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f\": rpc error: code = NotFound desc = could not find container \"a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f\": container with ID starting with a1eea1029b291eadd16793da90097645a8a9241d7764912b4498e06991e1847f not found: ID does not exist" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.494941 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.498305 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7b5b9c6757-jpdnk"] Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.634014 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-config\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.634673 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-client-ca\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.634992 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7faa616-a4d6-4131-8f3a-bbca90eeca33-serving-cert\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.635339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdx6g\" (UniqueName: \"kubernetes.io/projected/d7faa616-a4d6-4131-8f3a-bbca90eeca33-kube-api-access-bdx6g\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.736752 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7faa616-a4d6-4131-8f3a-bbca90eeca33-serving-cert\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.736902 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdx6g\" (UniqueName: \"kubernetes.io/projected/d7faa616-a4d6-4131-8f3a-bbca90eeca33-kube-api-access-bdx6g\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.736986 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-config\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.737089 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-client-ca\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.738713 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-config\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.738951 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7faa616-a4d6-4131-8f3a-bbca90eeca33-client-ca\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.742280 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7faa616-a4d6-4131-8f3a-bbca90eeca33-serving-cert\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.761620 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdx6g\" (UniqueName: \"kubernetes.io/projected/d7faa616-a4d6-4131-8f3a-bbca90eeca33-kube-api-access-bdx6g\") pod \"route-controller-manager-56c5f44854-dlnzr\" (UID: \"d7faa616-a4d6-4131-8f3a-bbca90eeca33\") " pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:46 crc kubenswrapper[4751]: I1002 11:04:46.787544 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:47 crc kubenswrapper[4751]: W1002 11:04:47.058520 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7faa616_a4d6_4131_8f3a_bbca90eeca33.slice/crio-4b3aec546c7dae9ea5905d01a6961d4b1e36661448a2be830db6926a59332a45 WatchSource:0}: Error finding container 4b3aec546c7dae9ea5905d01a6961d4b1e36661448a2be830db6926a59332a45: Status 404 returned error can't find the container with id 4b3aec546c7dae9ea5905d01a6961d4b1e36661448a2be830db6926a59332a45 Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.058968 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr"] Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.451439 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" event={"ID":"d7faa616-a4d6-4131-8f3a-bbca90eeca33","Type":"ContainerStarted","Data":"81db75dbb4adc17bc74c1b4e9968ac8ca42b42aac545958f644123c3afc53b12"} Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.452575 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" event={"ID":"d7faa616-a4d6-4131-8f3a-bbca90eeca33","Type":"ContainerStarted","Data":"4b3aec546c7dae9ea5905d01a6961d4b1e36661448a2be830db6926a59332a45"} Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.452625 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.471921 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" podStartSLOduration=3.471898678 podStartE2EDuration="3.471898678s" podCreationTimestamp="2025-10-02 11:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:04:47.468520469 +0000 UTC m=+769.522747009" watchObservedRunningTime="2025-10-02 11:04:47.471898678 +0000 UTC m=+769.526125138" Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.559255 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d417eb5f-e9ef-4ea8-aec0-ba7186157c69" path="/var/lib/kubelet/pods/d417eb5f-e9ef-4ea8-aec0-ba7186157c69/volumes" Oct 02 11:04:47 crc kubenswrapper[4751]: I1002 11:04:47.753135 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56c5f44854-dlnzr" Oct 02 11:04:48 crc kubenswrapper[4751]: I1002 11:04:48.281796 4751 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 11:05:01 crc kubenswrapper[4751]: I1002 11:05:01.507291 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:05:01 crc kubenswrapper[4751]: I1002 11:05:01.507836 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:05:01 crc kubenswrapper[4751]: I1002 11:05:01.507877 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:05:01 crc kubenswrapper[4751]: I1002 11:05:01.508507 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:05:01 crc kubenswrapper[4751]: I1002 11:05:01.508568 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed" gracePeriod=600 Oct 02 11:05:02 crc kubenswrapper[4751]: I1002 11:05:02.538390 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed" exitCode=0 Oct 02 11:05:02 crc kubenswrapper[4751]: I1002 11:05:02.538457 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed"} Oct 02 11:05:02 crc kubenswrapper[4751]: I1002 11:05:02.538983 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c"} Oct 02 11:05:02 crc kubenswrapper[4751]: I1002 11:05:02.539001 4751 scope.go:117] "RemoveContainer" containerID="8d06f28f8ce3bc680031594e6db466d5720cfa938bedd010643d9e75dc10097a" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.212225 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.224819 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.231123 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.397965 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.398101 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr446\" (UniqueName: \"kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.398244 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.499329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.499408 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr446\" (UniqueName: \"kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.499444 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.499862 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.499896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.535147 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr446\" (UniqueName: \"kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446\") pod \"community-operators-twz7r\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:54 crc kubenswrapper[4751]: I1002 11:05:54.558130 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:05:55 crc kubenswrapper[4751]: I1002 11:05:55.002402 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:05:55 crc kubenswrapper[4751]: I1002 11:05:55.872814 4751 generic.go:334] "Generic (PLEG): container finished" podID="54a8bda1-fe06-4159-b214-f1239ad57415" containerID="89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef" exitCode=0 Oct 02 11:05:55 crc kubenswrapper[4751]: I1002 11:05:55.872921 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerDied","Data":"89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef"} Oct 02 11:05:55 crc kubenswrapper[4751]: I1002 11:05:55.873134 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerStarted","Data":"c988fa7f63706dea055889ae80094cb56f2e5781f221482b1eb13cec311bab9f"} Oct 02 11:05:55 crc kubenswrapper[4751]: I1002 11:05:55.875862 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:05:56 crc kubenswrapper[4751]: I1002 11:05:56.883005 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerStarted","Data":"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58"} Oct 02 11:05:57 crc kubenswrapper[4751]: I1002 11:05:57.892383 4751 generic.go:334] "Generic (PLEG): container finished" podID="54a8bda1-fe06-4159-b214-f1239ad57415" containerID="c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58" exitCode=0 Oct 02 11:05:57 crc kubenswrapper[4751]: I1002 11:05:57.892448 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerDied","Data":"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58"} Oct 02 11:05:58 crc kubenswrapper[4751]: I1002 11:05:58.904026 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerStarted","Data":"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b"} Oct 02 11:05:58 crc kubenswrapper[4751]: I1002 11:05:58.930899 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-twz7r" podStartSLOduration=2.305572611 podStartE2EDuration="4.930866079s" podCreationTimestamp="2025-10-02 11:05:54 +0000 UTC" firstStartedPulling="2025-10-02 11:05:55.87537003 +0000 UTC m=+837.929596510" lastFinishedPulling="2025-10-02 11:05:58.500663528 +0000 UTC m=+840.554889978" observedRunningTime="2025-10-02 11:05:58.926417691 +0000 UTC m=+840.980644131" watchObservedRunningTime="2025-10-02 11:05:58.930866079 +0000 UTC m=+840.985092529" Oct 02 11:06:04 crc kubenswrapper[4751]: I1002 11:06:04.559474 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:04 crc kubenswrapper[4751]: I1002 11:06:04.560866 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:04 crc kubenswrapper[4751]: I1002 11:06:04.633220 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:04 crc kubenswrapper[4751]: I1002 11:06:04.997959 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:05 crc kubenswrapper[4751]: I1002 11:06:05.046559 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:06:06 crc kubenswrapper[4751]: I1002 11:06:06.954933 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-twz7r" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="registry-server" containerID="cri-o://b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b" gracePeriod=2 Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.343111 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.466151 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content\") pod \"54a8bda1-fe06-4159-b214-f1239ad57415\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.466344 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities\") pod \"54a8bda1-fe06-4159-b214-f1239ad57415\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.467354 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities" (OuterVolumeSpecName: "utilities") pod "54a8bda1-fe06-4159-b214-f1239ad57415" (UID: "54a8bda1-fe06-4159-b214-f1239ad57415"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.467479 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr446\" (UniqueName: \"kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446\") pod \"54a8bda1-fe06-4159-b214-f1239ad57415\" (UID: \"54a8bda1-fe06-4159-b214-f1239ad57415\") " Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.468493 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.472599 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446" (OuterVolumeSpecName: "kube-api-access-wr446") pod "54a8bda1-fe06-4159-b214-f1239ad57415" (UID: "54a8bda1-fe06-4159-b214-f1239ad57415"). InnerVolumeSpecName "kube-api-access-wr446". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.569523 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr446\" (UniqueName: \"kubernetes.io/projected/54a8bda1-fe06-4159-b214-f1239ad57415-kube-api-access-wr446\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.964772 4751 generic.go:334] "Generic (PLEG): container finished" podID="54a8bda1-fe06-4159-b214-f1239ad57415" containerID="b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b" exitCode=0 Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.964840 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerDied","Data":"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b"} Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.964864 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twz7r" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.964891 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twz7r" event={"ID":"54a8bda1-fe06-4159-b214-f1239ad57415","Type":"ContainerDied","Data":"c988fa7f63706dea055889ae80094cb56f2e5781f221482b1eb13cec311bab9f"} Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.964931 4751 scope.go:117] "RemoveContainer" containerID="b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b" Oct 02 11:06:07 crc kubenswrapper[4751]: I1002 11:06:07.984822 4751 scope.go:117] "RemoveContainer" containerID="c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.009623 4751 scope.go:117] "RemoveContainer" containerID="89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.081749 4751 scope.go:117] "RemoveContainer" containerID="b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b" Oct 02 11:06:08 crc kubenswrapper[4751]: E1002 11:06:08.082617 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b\": container with ID starting with b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b not found: ID does not exist" containerID="b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.082670 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b"} err="failed to get container status \"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b\": rpc error: code = NotFound desc = could not find container \"b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b\": container with ID starting with b2a6076ccf989761c7a85681fb915e7faf07408a16a8d40e240a183fe624cd7b not found: ID does not exist" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.082704 4751 scope.go:117] "RemoveContainer" containerID="c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58" Oct 02 11:06:08 crc kubenswrapper[4751]: E1002 11:06:08.082968 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58\": container with ID starting with c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58 not found: ID does not exist" containerID="c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.082991 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58"} err="failed to get container status \"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58\": rpc error: code = NotFound desc = could not find container \"c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58\": container with ID starting with c080533c2d7cbbaa38f647efbb8191419e5e64bbcb82b31c0e7117c3b93cdf58 not found: ID does not exist" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.083005 4751 scope.go:117] "RemoveContainer" containerID="89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef" Oct 02 11:06:08 crc kubenswrapper[4751]: E1002 11:06:08.083286 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef\": container with ID starting with 89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef not found: ID does not exist" containerID="89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.083348 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef"} err="failed to get container status \"89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef\": rpc error: code = NotFound desc = could not find container \"89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef\": container with ID starting with 89a33f121984337dd677cd49b8dd19f3a4083c4c253412ca6a619975a9ce00ef not found: ID does not exist" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.161269 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54a8bda1-fe06-4159-b214-f1239ad57415" (UID: "54a8bda1-fe06-4159-b214-f1239ad57415"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.176633 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a8bda1-fe06-4159-b214-f1239ad57415-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.290829 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:06:08 crc kubenswrapper[4751]: I1002 11:06:08.293438 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-twz7r"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.208377 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbzvj"] Oct 02 11:06:09 crc kubenswrapper[4751]: E1002 11:06:09.209009 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="extract-utilities" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.209034 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="extract-utilities" Oct 02 11:06:09 crc kubenswrapper[4751]: E1002 11:06:09.209066 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="extract-content" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.209078 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="extract-content" Oct 02 11:06:09 crc kubenswrapper[4751]: E1002 11:06:09.209102 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="registry-server" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.209115 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="registry-server" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.209292 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" containerName="registry-server" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.209816 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.211536 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.211765 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bjlh6" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.212048 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.219974 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dn5sh"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.221373 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dn5sh" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.223671 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jgzhb" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.237846 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-m8wxt"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.238602 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.240244 4751 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-c2pq4" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.254395 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-m8wxt"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.256710 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbzvj"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.280573 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dn5sh"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.289262 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk6dv\" (UniqueName: \"kubernetes.io/projected/aa39b082-35a9-4465-ae1f-8ef168d22ff4-kube-api-access-sk6dv\") pod \"cert-manager-webhook-5655c58dd6-m8wxt\" (UID: \"aa39b082-35a9-4465-ae1f-8ef168d22ff4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.289324 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf4xr\" (UniqueName: \"kubernetes.io/projected/cef5a193-4c48-4da3-9518-0dd2fafc905a-kube-api-access-gf4xr\") pod \"cert-manager-5b446d88c5-dn5sh\" (UID: \"cef5a193-4c48-4da3-9518-0dd2fafc905a\") " pod="cert-manager/cert-manager-5b446d88c5-dn5sh" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.289431 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktjdh\" (UniqueName: \"kubernetes.io/projected/b31d2144-af31-43e1-a625-2f558c8b21a9-kube-api-access-ktjdh\") pod \"cert-manager-cainjector-7f985d654d-cbzvj\" (UID: \"b31d2144-af31-43e1-a625-2f558c8b21a9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.390047 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk6dv\" (UniqueName: \"kubernetes.io/projected/aa39b082-35a9-4465-ae1f-8ef168d22ff4-kube-api-access-sk6dv\") pod \"cert-manager-webhook-5655c58dd6-m8wxt\" (UID: \"aa39b082-35a9-4465-ae1f-8ef168d22ff4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.390099 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf4xr\" (UniqueName: \"kubernetes.io/projected/cef5a193-4c48-4da3-9518-0dd2fafc905a-kube-api-access-gf4xr\") pod \"cert-manager-5b446d88c5-dn5sh\" (UID: \"cef5a193-4c48-4da3-9518-0dd2fafc905a\") " pod="cert-manager/cert-manager-5b446d88c5-dn5sh" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.390124 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktjdh\" (UniqueName: \"kubernetes.io/projected/b31d2144-af31-43e1-a625-2f558c8b21a9-kube-api-access-ktjdh\") pod \"cert-manager-cainjector-7f985d654d-cbzvj\" (UID: \"b31d2144-af31-43e1-a625-2f558c8b21a9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.405231 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk6dv\" (UniqueName: \"kubernetes.io/projected/aa39b082-35a9-4465-ae1f-8ef168d22ff4-kube-api-access-sk6dv\") pod \"cert-manager-webhook-5655c58dd6-m8wxt\" (UID: \"aa39b082-35a9-4465-ae1f-8ef168d22ff4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.405952 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf4xr\" (UniqueName: \"kubernetes.io/projected/cef5a193-4c48-4da3-9518-0dd2fafc905a-kube-api-access-gf4xr\") pod \"cert-manager-5b446d88c5-dn5sh\" (UID: \"cef5a193-4c48-4da3-9518-0dd2fafc905a\") " pod="cert-manager/cert-manager-5b446d88c5-dn5sh" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.407057 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktjdh\" (UniqueName: \"kubernetes.io/projected/b31d2144-af31-43e1-a625-2f558c8b21a9-kube-api-access-ktjdh\") pod \"cert-manager-cainjector-7f985d654d-cbzvj\" (UID: \"b31d2144-af31-43e1-a625-2f558c8b21a9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.535988 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.548356 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dn5sh" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.556318 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.559439 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a8bda1-fe06-4159-b214-f1239ad57415" path="/var/lib/kubelet/pods/54a8bda1-fe06-4159-b214-f1239ad57415/volumes" Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.761033 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cbzvj"] Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.809242 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-m8wxt"] Oct 02 11:06:09 crc kubenswrapper[4751]: W1002 11:06:09.813209 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa39b082_35a9_4465_ae1f_8ef168d22ff4.slice/crio-b823a464641701f7e96c011095ce81c974a4317184320e53030b52607fc2d391 WatchSource:0}: Error finding container b823a464641701f7e96c011095ce81c974a4317184320e53030b52607fc2d391: Status 404 returned error can't find the container with id b823a464641701f7e96c011095ce81c974a4317184320e53030b52607fc2d391 Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.827125 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dn5sh"] Oct 02 11:06:09 crc kubenswrapper[4751]: W1002 11:06:09.831857 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcef5a193_4c48_4da3_9518_0dd2fafc905a.slice/crio-67324241cc3b6bb674020c2aa8b6e0a88a1bd212c4994323460a6919457b692b WatchSource:0}: Error finding container 67324241cc3b6bb674020c2aa8b6e0a88a1bd212c4994323460a6919457b692b: Status 404 returned error can't find the container with id 67324241cc3b6bb674020c2aa8b6e0a88a1bd212c4994323460a6919457b692b Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.976918 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" event={"ID":"b31d2144-af31-43e1-a625-2f558c8b21a9","Type":"ContainerStarted","Data":"deb66b161135619e2a5ac23a41b60ab8fb1f22ba05b88c292c7a98aa791cd2ba"} Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.978093 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" event={"ID":"aa39b082-35a9-4465-ae1f-8ef168d22ff4","Type":"ContainerStarted","Data":"b823a464641701f7e96c011095ce81c974a4317184320e53030b52607fc2d391"} Oct 02 11:06:09 crc kubenswrapper[4751]: I1002 11:06:09.979324 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dn5sh" event={"ID":"cef5a193-4c48-4da3-9518-0dd2fafc905a","Type":"ContainerStarted","Data":"67324241cc3b6bb674020c2aa8b6e0a88a1bd212c4994323460a6919457b692b"} Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.024733 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dn5sh" event={"ID":"cef5a193-4c48-4da3-9518-0dd2fafc905a","Type":"ContainerStarted","Data":"165399bdf334ddcd2a44cef52a2c982a7be4235ce0d7f84006d35fd726ab7087"} Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.028322 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" event={"ID":"b31d2144-af31-43e1-a625-2f558c8b21a9","Type":"ContainerStarted","Data":"ca857d251eb5c374b1f77bdecd2574053df98c99312213a564e81202e99dc477"} Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.030662 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" event={"ID":"aa39b082-35a9-4465-ae1f-8ef168d22ff4","Type":"ContainerStarted","Data":"08f372fca8ca07e4707b2d999569549b62734e9b117b3f4b6314bbb862f39636"} Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.030848 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.043969 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-dn5sh" podStartSLOduration=1.6523124359999999 podStartE2EDuration="5.043945358s" podCreationTimestamp="2025-10-02 11:06:09 +0000 UTC" firstStartedPulling="2025-10-02 11:06:09.833761461 +0000 UTC m=+851.887987911" lastFinishedPulling="2025-10-02 11:06:13.225394353 +0000 UTC m=+855.279620833" observedRunningTime="2025-10-02 11:06:14.04328157 +0000 UTC m=+856.097508020" watchObservedRunningTime="2025-10-02 11:06:14.043945358 +0000 UTC m=+856.098171808" Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.068498 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" podStartSLOduration=1.676154285 podStartE2EDuration="5.068480785s" podCreationTimestamp="2025-10-02 11:06:09 +0000 UTC" firstStartedPulling="2025-10-02 11:06:09.819752101 +0000 UTC m=+851.873978551" lastFinishedPulling="2025-10-02 11:06:13.212078591 +0000 UTC m=+855.266305051" observedRunningTime="2025-10-02 11:06:14.067336604 +0000 UTC m=+856.121563044" watchObservedRunningTime="2025-10-02 11:06:14.068480785 +0000 UTC m=+856.122707235" Oct 02 11:06:14 crc kubenswrapper[4751]: I1002 11:06:14.086102 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-cbzvj" podStartSLOduration=1.634695408 podStartE2EDuration="5.086080563s" podCreationTimestamp="2025-10-02 11:06:09 +0000 UTC" firstStartedPulling="2025-10-02 11:06:09.773681369 +0000 UTC m=+851.827907819" lastFinishedPulling="2025-10-02 11:06:13.225066524 +0000 UTC m=+855.279292974" observedRunningTime="2025-10-02 11:06:14.083229976 +0000 UTC m=+856.137456426" watchObservedRunningTime="2025-10-02 11:06:14.086080563 +0000 UTC m=+856.140307043" Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.562532 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-m8wxt" Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.880750 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pccq9"] Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881274 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-controller" containerID="cri-o://2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881409 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="nbdb" containerID="cri-o://8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881447 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881493 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="sbdb" containerID="cri-o://1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881566 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-node" containerID="cri-o://3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881606 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-acl-logging" containerID="cri-o://f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.881610 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="northd" containerID="cri-o://cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" gracePeriod=30 Oct 02 11:06:19 crc kubenswrapper[4751]: I1002 11:06:19.936436 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" containerID="cri-o://e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" gracePeriod=30 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.078864 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovnkube-controller/3.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.082755 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-acl-logging/0.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.083497 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-controller/0.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084249 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" exitCode=0 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084278 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" exitCode=0 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084292 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" exitCode=0 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084303 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" exitCode=143 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084314 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" exitCode=143 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084339 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084396 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084414 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084425 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084436 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.084457 4751 scope.go:117] "RemoveContainer" containerID="a7517b702dec95990881912c597bb79ca9bccd2da5b061d28b46d469a9fe429d" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.087903 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/2.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.088326 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/1.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.088379 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e66e701-8712-4399-b2a8-b42e8cf716a3" containerID="4c0bf2e16e4a69962d135faba19b1da383c6457fc10f7ded584b28ab1d7ab51a" exitCode=2 Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.088414 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerDied","Data":"4c0bf2e16e4a69962d135faba19b1da383c6457fc10f7ded584b28ab1d7ab51a"} Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.088938 4751 scope.go:117] "RemoveContainer" containerID="4c0bf2e16e4a69962d135faba19b1da383c6457fc10f7ded584b28ab1d7ab51a" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.224715 4751 scope.go:117] "RemoveContainer" containerID="ce9b25f6a7cb81076494c70a103356fdbe9e8cbaeeee14a13706beab369df2f2" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.240864 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-acl-logging/0.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.241667 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-controller/0.log" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.242289 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.295712 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hd45s"] Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.295946 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.295967 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.295981 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.295989 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.295998 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296006 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296019 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296027 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296036 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296043 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296054 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-acl-logging" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296062 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-acl-logging" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296070 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kubecfg-setup" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296077 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kubecfg-setup" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296085 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-node" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296092 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-node" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296100 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="nbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296106 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="nbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296113 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="northd" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296120 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="northd" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296129 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="sbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296135 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="sbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296146 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296151 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 11:06:20 crc kubenswrapper[4751]: E1002 11:06:20.296161 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296167 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296318 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="sbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296332 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296340 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296352 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296361 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296368 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296376 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="kube-rbac-proxy-node" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296384 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="northd" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296392 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-acl-logging" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296399 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="nbdb" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296407 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovn-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.296563 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerName="ovnkube-controller" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.298926 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441248 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441309 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441342 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441377 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgl42\" (UniqueName: \"kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441412 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441440 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441458 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441462 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441494 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441485 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441535 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash" (OuterVolumeSpecName: "host-slash") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441586 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441626 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441627 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441656 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441694 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441697 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441715 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441734 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441772 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441803 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441837 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441872 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441916 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441970 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441869 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log" (OuterVolumeSpecName: "node-log") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441897 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441916 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442006 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442043 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442063 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442084 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442104 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin\") pod \"37ba8a11-317d-406b-8628-fcd43b164dbd\" (UID: \"37ba8a11-317d-406b-8628-fcd43b164dbd\") " Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442197 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442302 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-kubelet\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442335 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-env-overrides\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442366 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-netns\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442392 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-systemd-units\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442428 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-script-lib\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442486 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-bin\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442522 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-var-lib-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442540 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442549 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.441461 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket" (OuterVolumeSpecName: "log-socket") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442621 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-log-socket\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442659 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-systemd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442715 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47acac24-e8cf-4984-9655-521f5c67d2df-ovn-node-metrics-cert\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442742 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442766 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-etc-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442794 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442831 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-netd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh5pz\" (UniqueName: \"kubernetes.io/projected/47acac24-e8cf-4984-9655-521f5c67d2df-kube-api-access-sh5pz\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442890 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-slash\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442932 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-ovn\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-config\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.442979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-node-log\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443138 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443186 4751 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443224 4751 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443240 4751 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443253 4751 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443264 4751 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443278 4751 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443292 4751 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443303 4751 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443316 4751 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37ba8a11-317d-406b-8628-fcd43b164dbd-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443328 4751 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443340 4751 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443354 4751 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443366 4751 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443378 4751 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443389 4751 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.443402 4751 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.448519 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42" (OuterVolumeSpecName: "kube-api-access-lgl42") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "kube-api-access-lgl42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.449264 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.455537 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "37ba8a11-317d-406b-8628-fcd43b164dbd" (UID: "37ba8a11-317d-406b-8628-fcd43b164dbd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.544815 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.544994 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-log-socket\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545256 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-log-socket\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545381 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-systemd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545430 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47acac24-e8cf-4984-9655-521f5c67d2df-ovn-node-metrics-cert\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545462 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545491 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-etc-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545524 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545583 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-netd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545609 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-slash\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545633 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh5pz\" (UniqueName: \"kubernetes.io/projected/47acac24-e8cf-4984-9655-521f5c67d2df-kube-api-access-sh5pz\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545657 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545696 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-ovn\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545720 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-netd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545672 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-ovn\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545689 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-run-systemd\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545743 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-etc-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545766 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-config\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545890 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-node-log\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545980 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-kubelet\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546013 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-env-overrides\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545766 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.545784 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-slash\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546154 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-node-log\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546641 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-config\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546762 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-kubelet\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546831 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-netns\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546869 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-run-netns\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.546942 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-systemd-units\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547011 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-script-lib\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547032 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-systemd-units\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547043 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-bin\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547125 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-var-lib-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547247 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-var-lib-openvswitch\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547308 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgl42\" (UniqueName: \"kubernetes.io/projected/37ba8a11-317d-406b-8628-fcd43b164dbd-kube-api-access-lgl42\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547332 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/47acac24-e8cf-4984-9655-521f5c67d2df-host-cni-bin\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547344 4751 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37ba8a11-317d-406b-8628-fcd43b164dbd-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547355 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37ba8a11-317d-406b-8628-fcd43b164dbd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.547683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-env-overrides\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.548632 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/47acac24-e8cf-4984-9655-521f5c67d2df-ovnkube-script-lib\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.548846 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47acac24-e8cf-4984-9655-521f5c67d2df-ovn-node-metrics-cert\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.567288 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh5pz\" (UniqueName: \"kubernetes.io/projected/47acac24-e8cf-4984-9655-521f5c67d2df-kube-api-access-sh5pz\") pod \"ovnkube-node-hd45s\" (UID: \"47acac24-e8cf-4984-9655-521f5c67d2df\") " pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: I1002 11:06:20.615455 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:20 crc kubenswrapper[4751]: W1002 11:06:20.633639 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47acac24_e8cf_4984_9655_521f5c67d2df.slice/crio-c35911629f8ad9d2283cadb24e7f57cc28594e18a0cd4d50a9561562971c6edd WatchSource:0}: Error finding container c35911629f8ad9d2283cadb24e7f57cc28594e18a0cd4d50a9561562971c6edd: Status 404 returned error can't find the container with id c35911629f8ad9d2283cadb24e7f57cc28594e18a0cd4d50a9561562971c6edd Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.100230 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8vj4p_5e66e701-8712-4399-b2a8-b42e8cf716a3/kube-multus/2.log" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.100392 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8vj4p" event={"ID":"5e66e701-8712-4399-b2a8-b42e8cf716a3","Type":"ContainerStarted","Data":"6484fef18ea04c53e744c6e3120d35099f58119761c740cebeb75e214cd988e9"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.103470 4751 generic.go:334] "Generic (PLEG): container finished" podID="47acac24-e8cf-4984-9655-521f5c67d2df" containerID="d26b5a0d96c61d322cc227ff318e884a5fc1d4a1a2461034c5b0b517e4efc19f" exitCode=0 Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.103610 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerDied","Data":"d26b5a0d96c61d322cc227ff318e884a5fc1d4a1a2461034c5b0b517e4efc19f"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.104640 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"c35911629f8ad9d2283cadb24e7f57cc28594e18a0cd4d50a9561562971c6edd"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.114402 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-acl-logging/0.log" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115039 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pccq9_37ba8a11-317d-406b-8628-fcd43b164dbd/ovn-controller/0.log" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115592 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" exitCode=0 Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115618 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" exitCode=0 Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115643 4751 generic.go:334] "Generic (PLEG): container finished" podID="37ba8a11-317d-406b-8628-fcd43b164dbd" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" exitCode=0 Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115673 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115698 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115711 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115722 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" event={"ID":"37ba8a11-317d-406b-8628-fcd43b164dbd","Type":"ContainerDied","Data":"14a180816373f064036945a715f84b6144ed14b8a988f0604480e57cee77451a"} Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115742 4751 scope.go:117] "RemoveContainer" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.115769 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pccq9" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.140827 4751 scope.go:117] "RemoveContainer" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.159102 4751 scope.go:117] "RemoveContainer" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.179881 4751 scope.go:117] "RemoveContainer" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.194955 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pccq9"] Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.201620 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pccq9"] Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.211907 4751 scope.go:117] "RemoveContainer" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.230831 4751 scope.go:117] "RemoveContainer" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.250156 4751 scope.go:117] "RemoveContainer" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.270642 4751 scope.go:117] "RemoveContainer" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.284953 4751 scope.go:117] "RemoveContainer" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.301510 4751 scope.go:117] "RemoveContainer" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.305281 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": container with ID starting with e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3 not found: ID does not exist" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.305322 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3"} err="failed to get container status \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": rpc error: code = NotFound desc = could not find container \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": container with ID starting with e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.305349 4751 scope.go:117] "RemoveContainer" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.308979 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": container with ID starting with 1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c not found: ID does not exist" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.309013 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c"} err="failed to get container status \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": rpc error: code = NotFound desc = could not find container \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": container with ID starting with 1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.309037 4751 scope.go:117] "RemoveContainer" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.309685 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": container with ID starting with 8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143 not found: ID does not exist" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.309735 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143"} err="failed to get container status \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": rpc error: code = NotFound desc = could not find container \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": container with ID starting with 8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.309776 4751 scope.go:117] "RemoveContainer" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.310119 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": container with ID starting with cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a not found: ID does not exist" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310146 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a"} err="failed to get container status \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": rpc error: code = NotFound desc = could not find container \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": container with ID starting with cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310163 4751 scope.go:117] "RemoveContainer" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.310396 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": container with ID starting with 9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3 not found: ID does not exist" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310427 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3"} err="failed to get container status \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": rpc error: code = NotFound desc = could not find container \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": container with ID starting with 9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310441 4751 scope.go:117] "RemoveContainer" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.310677 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": container with ID starting with 3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217 not found: ID does not exist" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310702 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217"} err="failed to get container status \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": rpc error: code = NotFound desc = could not find container \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": container with ID starting with 3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.310715 4751 scope.go:117] "RemoveContainer" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.310982 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": container with ID starting with f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded not found: ID does not exist" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311012 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded"} err="failed to get container status \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": rpc error: code = NotFound desc = could not find container \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": container with ID starting with f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311026 4751 scope.go:117] "RemoveContainer" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.311288 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": container with ID starting with 2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf not found: ID does not exist" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311307 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf"} err="failed to get container status \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": rpc error: code = NotFound desc = could not find container \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": container with ID starting with 2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311319 4751 scope.go:117] "RemoveContainer" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" Oct 02 11:06:21 crc kubenswrapper[4751]: E1002 11:06:21.311502 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": container with ID starting with aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af not found: ID does not exist" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311540 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af"} err="failed to get container status \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": rpc error: code = NotFound desc = could not find container \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": container with ID starting with aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311558 4751 scope.go:117] "RemoveContainer" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311828 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3"} err="failed to get container status \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": rpc error: code = NotFound desc = could not find container \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": container with ID starting with e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.311849 4751 scope.go:117] "RemoveContainer" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.312310 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c"} err="failed to get container status \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": rpc error: code = NotFound desc = could not find container \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": container with ID starting with 1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.312423 4751 scope.go:117] "RemoveContainer" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.312862 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143"} err="failed to get container status \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": rpc error: code = NotFound desc = could not find container \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": container with ID starting with 8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.312897 4751 scope.go:117] "RemoveContainer" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313112 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a"} err="failed to get container status \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": rpc error: code = NotFound desc = could not find container \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": container with ID starting with cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313139 4751 scope.go:117] "RemoveContainer" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313392 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3"} err="failed to get container status \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": rpc error: code = NotFound desc = could not find container \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": container with ID starting with 9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313410 4751 scope.go:117] "RemoveContainer" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313630 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217"} err="failed to get container status \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": rpc error: code = NotFound desc = could not find container \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": container with ID starting with 3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.313660 4751 scope.go:117] "RemoveContainer" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.314021 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded"} err="failed to get container status \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": rpc error: code = NotFound desc = could not find container \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": container with ID starting with f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.314039 4751 scope.go:117] "RemoveContainer" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.314263 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf"} err="failed to get container status \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": rpc error: code = NotFound desc = could not find container \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": container with ID starting with 2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.314281 4751 scope.go:117] "RemoveContainer" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317064 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af"} err="failed to get container status \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": rpc error: code = NotFound desc = could not find container \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": container with ID starting with aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317107 4751 scope.go:117] "RemoveContainer" containerID="e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317452 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3"} err="failed to get container status \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": rpc error: code = NotFound desc = could not find container \"e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3\": container with ID starting with e32aabfb1350c29f5d2d3bd26c6e61627319596729d59271c9a43951b9480ba3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317472 4751 scope.go:117] "RemoveContainer" containerID="1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317879 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c"} err="failed to get container status \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": rpc error: code = NotFound desc = could not find container \"1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c\": container with ID starting with 1531247640feccee4e8a0b0c37cdbabced900308559862f8a555c0f738cc4e4c not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.317912 4751 scope.go:117] "RemoveContainer" containerID="8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318095 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143"} err="failed to get container status \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": rpc error: code = NotFound desc = could not find container \"8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143\": container with ID starting with 8b009f944d8948ae8ff63ecd9e61e4d44977f6a9bbe498cf3ef8e5577b7f7143 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318126 4751 scope.go:117] "RemoveContainer" containerID="cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318306 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a"} err="failed to get container status \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": rpc error: code = NotFound desc = could not find container \"cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a\": container with ID starting with cf8497f9ecf0be59de72803ab718c02e1f467d4dcee63dab7d0924ff527b4a1a not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318324 4751 scope.go:117] "RemoveContainer" containerID="9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318865 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3"} err="failed to get container status \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": rpc error: code = NotFound desc = could not find container \"9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3\": container with ID starting with 9d40ec56528d33afbb005d13314234bc652ba4b41708383503413b5acebeabf3 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.318889 4751 scope.go:117] "RemoveContainer" containerID="3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322027 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217"} err="failed to get container status \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": rpc error: code = NotFound desc = could not find container \"3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217\": container with ID starting with 3bdc21febf72e4509a8da1306b95f9a2ca89abb8b96d9d7ba34087b8351c7217 not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322053 4751 scope.go:117] "RemoveContainer" containerID="f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322437 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded"} err="failed to get container status \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": rpc error: code = NotFound desc = could not find container \"f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded\": container with ID starting with f9384c92c8e2a3ce68af504db6a8765e9522853be6987b6aeac108f84dbe9ded not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322498 4751 scope.go:117] "RemoveContainer" containerID="2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322870 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf"} err="failed to get container status \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": rpc error: code = NotFound desc = could not find container \"2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf\": container with ID starting with 2f543f1d819c12ef0543ffdeb4187676c1038ee483b94bf1c7140170fec5f3cf not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.322899 4751 scope.go:117] "RemoveContainer" containerID="aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.323219 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af"} err="failed to get container status \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": rpc error: code = NotFound desc = could not find container \"aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af\": container with ID starting with aeb4c1b0dca27d0eabf7e93e463615e3796ce55ac24ce369e05ce373017846af not found: ID does not exist" Oct 02 11:06:21 crc kubenswrapper[4751]: I1002 11:06:21.558433 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ba8a11-317d-406b-8628-fcd43b164dbd" path="/var/lib/kubelet/pods/37ba8a11-317d-406b-8628-fcd43b164dbd/volumes" Oct 02 11:06:22 crc kubenswrapper[4751]: I1002 11:06:22.127438 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"a0a1108e40161b5892aac2981a0c90f84c7f8748a371ae7b8fd82c257c3f9fe8"} Oct 02 11:06:22 crc kubenswrapper[4751]: I1002 11:06:22.127806 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"eefb3b3278961381db4ff626ae94b3b86a61d3e240b6a0cdaf7f9936a91985fe"} Oct 02 11:06:22 crc kubenswrapper[4751]: I1002 11:06:22.127821 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"2bd97d16fc5c19d75952d09fd55b43b4cf779345eec09f2e73c2aedfddd81b92"} Oct 02 11:06:22 crc kubenswrapper[4751]: I1002 11:06:22.127833 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"0cbaf682d88bd1349314a597fe1e6f08a4ffe0f7cc4540b3f481326f5c7e4a30"} Oct 02 11:06:22 crc kubenswrapper[4751]: I1002 11:06:22.127873 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"054e812dae8c58956f945f419a08f8faf59fa037851be2d3c101c316f069471a"} Oct 02 11:06:23 crc kubenswrapper[4751]: I1002 11:06:23.138142 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"a9b2068f768e91cf1a399007d339bd259da0abce33e869f9ffd034b988c13437"} Oct 02 11:06:25 crc kubenswrapper[4751]: I1002 11:06:25.156719 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"109aeb908a10ba899e661c8972e1e3cf219937bac481881aa6fbb2bbef0ac19a"} Oct 02 11:06:28 crc kubenswrapper[4751]: I1002 11:06:28.181417 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" event={"ID":"47acac24-e8cf-4984-9655-521f5c67d2df","Type":"ContainerStarted","Data":"f0e728dd674cab986b34ad054604fbf442cbf85bd4dc7057d5a7714a7e0745a0"} Oct 02 11:06:28 crc kubenswrapper[4751]: I1002 11:06:28.182000 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:28 crc kubenswrapper[4751]: I1002 11:06:28.225674 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:28 crc kubenswrapper[4751]: I1002 11:06:28.226481 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" podStartSLOduration=8.226466507 podStartE2EDuration="8.226466507s" podCreationTimestamp="2025-10-02 11:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:06:28.224327699 +0000 UTC m=+870.278554189" watchObservedRunningTime="2025-10-02 11:06:28.226466507 +0000 UTC m=+870.280692957" Oct 02 11:06:29 crc kubenswrapper[4751]: I1002 11:06:29.188814 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:29 crc kubenswrapper[4751]: I1002 11:06:29.188902 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:29 crc kubenswrapper[4751]: I1002 11:06:29.232406 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.513395 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.515995 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.530372 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.636652 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.636930 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.636998 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lslwk\" (UniqueName: \"kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.738588 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lslwk\" (UniqueName: \"kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.738736 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.738757 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.739392 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.739484 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.773547 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lslwk\" (UniqueName: \"kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk\") pod \"redhat-operators-rbp6d\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:35 crc kubenswrapper[4751]: I1002 11:06:35.836641 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:36 crc kubenswrapper[4751]: I1002 11:06:36.038816 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:36 crc kubenswrapper[4751]: W1002 11:06:36.046930 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4df81173_f48b_4f0c_9ea2_2f8bd7df0cb0.slice/crio-06e9f134426ecf999452eb677769369bf0c78f0f150f24984c12428005328a55 WatchSource:0}: Error finding container 06e9f134426ecf999452eb677769369bf0c78f0f150f24984c12428005328a55: Status 404 returned error can't find the container with id 06e9f134426ecf999452eb677769369bf0c78f0f150f24984c12428005328a55 Oct 02 11:06:36 crc kubenswrapper[4751]: I1002 11:06:36.229852 4751 generic.go:334] "Generic (PLEG): container finished" podID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerID="30f754fdbb4637a025aaf7df08cbf4db1e05b094dc25338b129e93f2aee5846e" exitCode=0 Oct 02 11:06:36 crc kubenswrapper[4751]: I1002 11:06:36.229908 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerDied","Data":"30f754fdbb4637a025aaf7df08cbf4db1e05b094dc25338b129e93f2aee5846e"} Oct 02 11:06:36 crc kubenswrapper[4751]: I1002 11:06:36.229935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerStarted","Data":"06e9f134426ecf999452eb677769369bf0c78f0f150f24984c12428005328a55"} Oct 02 11:06:38 crc kubenswrapper[4751]: I1002 11:06:38.891805 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:38 crc kubenswrapper[4751]: I1002 11:06:38.893219 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:38 crc kubenswrapper[4751]: I1002 11:06:38.914302 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.082211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.082278 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gr8c\" (UniqueName: \"kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.082344 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.183478 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.183538 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gr8c\" (UniqueName: \"kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.183578 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.184219 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.184254 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.207436 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gr8c\" (UniqueName: \"kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c\") pod \"certified-operators-4cdnw\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.248017 4751 generic.go:334] "Generic (PLEG): container finished" podID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerID="859782b004cf021dd7d31d85231618b21dd2d825bf87c1f35ef1d0d648c1398a" exitCode=0 Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.248058 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerDied","Data":"859782b004cf021dd7d31d85231618b21dd2d825bf87c1f35ef1d0d648c1398a"} Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.508303 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:39 crc kubenswrapper[4751]: I1002 11:06:39.717585 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:39 crc kubenswrapper[4751]: W1002 11:06:39.726106 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf9e0fb5_080e_4358_ae7c_de730482a384.slice/crio-271b52000963f68ffb2304510b812ef8d88968688becdc357b884b97d5b842c3 WatchSource:0}: Error finding container 271b52000963f68ffb2304510b812ef8d88968688becdc357b884b97d5b842c3: Status 404 returned error can't find the container with id 271b52000963f68ffb2304510b812ef8d88968688becdc357b884b97d5b842c3 Oct 02 11:06:40 crc kubenswrapper[4751]: I1002 11:06:40.257467 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerStarted","Data":"b2a3539e6c0b1d4265c7e61d96f5de32144505d37cfbe08cc0f5e04449da3abb"} Oct 02 11:06:40 crc kubenswrapper[4751]: I1002 11:06:40.261023 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerID="76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761" exitCode=0 Oct 02 11:06:40 crc kubenswrapper[4751]: I1002 11:06:40.261079 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerDied","Data":"76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761"} Oct 02 11:06:40 crc kubenswrapper[4751]: I1002 11:06:40.261105 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerStarted","Data":"271b52000963f68ffb2304510b812ef8d88968688becdc357b884b97d5b842c3"} Oct 02 11:06:40 crc kubenswrapper[4751]: I1002 11:06:40.278814 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rbp6d" podStartSLOduration=1.7620387499999999 podStartE2EDuration="5.278794261s" podCreationTimestamp="2025-10-02 11:06:35 +0000 UTC" firstStartedPulling="2025-10-02 11:06:36.231078138 +0000 UTC m=+878.285304588" lastFinishedPulling="2025-10-02 11:06:39.747833649 +0000 UTC m=+881.802060099" observedRunningTime="2025-10-02 11:06:40.275019748 +0000 UTC m=+882.329246208" watchObservedRunningTime="2025-10-02 11:06:40.278794261 +0000 UTC m=+882.333020721" Oct 02 11:06:41 crc kubenswrapper[4751]: I1002 11:06:41.267618 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerStarted","Data":"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea"} Oct 02 11:06:41 crc kubenswrapper[4751]: E1002 11:06:41.398035 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf9e0fb5_080e_4358_ae7c_de730482a384.slice/crio-838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf9e0fb5_080e_4358_ae7c_de730482a384.slice/crio-conmon-838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:06:42 crc kubenswrapper[4751]: I1002 11:06:42.274978 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerID="838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea" exitCode=0 Oct 02 11:06:42 crc kubenswrapper[4751]: I1002 11:06:42.275056 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerDied","Data":"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea"} Oct 02 11:06:43 crc kubenswrapper[4751]: I1002 11:06:43.284782 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerStarted","Data":"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd"} Oct 02 11:06:43 crc kubenswrapper[4751]: I1002 11:06:43.304304 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4cdnw" podStartSLOduration=2.846751239 podStartE2EDuration="5.304285147s" podCreationTimestamp="2025-10-02 11:06:38 +0000 UTC" firstStartedPulling="2025-10-02 11:06:40.263116355 +0000 UTC m=+882.317342815" lastFinishedPulling="2025-10-02 11:06:42.720650273 +0000 UTC m=+884.774876723" observedRunningTime="2025-10-02 11:06:43.302229251 +0000 UTC m=+885.356455781" watchObservedRunningTime="2025-10-02 11:06:43.304285147 +0000 UTC m=+885.358511597" Oct 02 11:06:45 crc kubenswrapper[4751]: I1002 11:06:45.837403 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:45 crc kubenswrapper[4751]: I1002 11:06:45.837475 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:45 crc kubenswrapper[4751]: I1002 11:06:45.887377 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:46 crc kubenswrapper[4751]: I1002 11:06:46.373865 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:47 crc kubenswrapper[4751]: I1002 11:06:47.082987 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:48 crc kubenswrapper[4751]: I1002 11:06:48.318428 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rbp6d" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="registry-server" containerID="cri-o://b2a3539e6c0b1d4265c7e61d96f5de32144505d37cfbe08cc0f5e04449da3abb" gracePeriod=2 Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.330248 4751 generic.go:334] "Generic (PLEG): container finished" podID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerID="b2a3539e6c0b1d4265c7e61d96f5de32144505d37cfbe08cc0f5e04449da3abb" exitCode=0 Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.330320 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerDied","Data":"b2a3539e6c0b1d4265c7e61d96f5de32144505d37cfbe08cc0f5e04449da3abb"} Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.508975 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.509069 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.578955 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:49 crc kubenswrapper[4751]: I1002 11:06:49.891637 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.022616 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content\") pod \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.022804 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities\") pod \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.022887 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lslwk\" (UniqueName: \"kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk\") pod \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\" (UID: \"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0\") " Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.024058 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities" (OuterVolumeSpecName: "utilities") pod "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" (UID: "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.031572 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk" (OuterVolumeSpecName: "kube-api-access-lslwk") pod "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" (UID: "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0"). InnerVolumeSpecName "kube-api-access-lslwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.124048 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.124108 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lslwk\" (UniqueName: \"kubernetes.io/projected/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-kube-api-access-lslwk\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.145384 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" (UID: "4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.224906 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.339355 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rbp6d" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.339359 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rbp6d" event={"ID":"4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0","Type":"ContainerDied","Data":"06e9f134426ecf999452eb677769369bf0c78f0f150f24984c12428005328a55"} Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.339452 4751 scope.go:117] "RemoveContainer" containerID="b2a3539e6c0b1d4265c7e61d96f5de32144505d37cfbe08cc0f5e04449da3abb" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.369482 4751 scope.go:117] "RemoveContainer" containerID="859782b004cf021dd7d31d85231618b21dd2d825bf87c1f35ef1d0d648c1398a" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.376451 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.381985 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rbp6d"] Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.389835 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.395900 4751 scope.go:117] "RemoveContainer" containerID="30f754fdbb4637a025aaf7df08cbf4db1e05b094dc25338b129e93f2aee5846e" Oct 02 11:06:50 crc kubenswrapper[4751]: I1002 11:06:50.640357 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hd45s" Oct 02 11:06:51 crc kubenswrapper[4751]: I1002 11:06:51.562337 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" path="/var/lib/kubelet/pods/4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0/volumes" Oct 02 11:06:52 crc kubenswrapper[4751]: I1002 11:06:52.679790 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:52 crc kubenswrapper[4751]: I1002 11:06:52.680034 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4cdnw" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="registry-server" containerID="cri-o://89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd" gracePeriod=2 Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.043306 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.162583 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content\") pod \"bf9e0fb5-080e-4358-ae7c-de730482a384\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.162651 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gr8c\" (UniqueName: \"kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c\") pod \"bf9e0fb5-080e-4358-ae7c-de730482a384\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.162707 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities\") pod \"bf9e0fb5-080e-4358-ae7c-de730482a384\" (UID: \"bf9e0fb5-080e-4358-ae7c-de730482a384\") " Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.163719 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities" (OuterVolumeSpecName: "utilities") pod "bf9e0fb5-080e-4358-ae7c-de730482a384" (UID: "bf9e0fb5-080e-4358-ae7c-de730482a384"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.171061 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c" (OuterVolumeSpecName: "kube-api-access-2gr8c") pod "bf9e0fb5-080e-4358-ae7c-de730482a384" (UID: "bf9e0fb5-080e-4358-ae7c-de730482a384"). InnerVolumeSpecName "kube-api-access-2gr8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.264295 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.264343 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gr8c\" (UniqueName: \"kubernetes.io/projected/bf9e0fb5-080e-4358-ae7c-de730482a384-kube-api-access-2gr8c\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.363313 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerID="89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd" exitCode=0 Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.363407 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cdnw" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.363430 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerDied","Data":"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd"} Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.364458 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cdnw" event={"ID":"bf9e0fb5-080e-4358-ae7c-de730482a384","Type":"ContainerDied","Data":"271b52000963f68ffb2304510b812ef8d88968688becdc357b884b97d5b842c3"} Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.364528 4751 scope.go:117] "RemoveContainer" containerID="89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.390485 4751 scope.go:117] "RemoveContainer" containerID="838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.408149 4751 scope.go:117] "RemoveContainer" containerID="76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.451104 4751 scope.go:117] "RemoveContainer" containerID="89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd" Oct 02 11:06:53 crc kubenswrapper[4751]: E1002 11:06:53.451785 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd\": container with ID starting with 89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd not found: ID does not exist" containerID="89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.451836 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd"} err="failed to get container status \"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd\": rpc error: code = NotFound desc = could not find container \"89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd\": container with ID starting with 89840cc9f2973e35541c802bfa7997471bc35a24fbf06dd5b2d83510e919b0cd not found: ID does not exist" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.451865 4751 scope.go:117] "RemoveContainer" containerID="838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea" Oct 02 11:06:53 crc kubenswrapper[4751]: E1002 11:06:53.452422 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea\": container with ID starting with 838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea not found: ID does not exist" containerID="838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.452497 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea"} err="failed to get container status \"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea\": rpc error: code = NotFound desc = could not find container \"838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea\": container with ID starting with 838efb917b24d527c5c772386c57a604cff8c19b97b61f27bfdc3202c0739fea not found: ID does not exist" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.452530 4751 scope.go:117] "RemoveContainer" containerID="76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761" Oct 02 11:06:53 crc kubenswrapper[4751]: E1002 11:06:53.453035 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761\": container with ID starting with 76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761 not found: ID does not exist" containerID="76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.453084 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761"} err="failed to get container status \"76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761\": rpc error: code = NotFound desc = could not find container \"76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761\": container with ID starting with 76f3d420c74b1d50435e368804fd839108ea6a68e3b800e6560489b32463f761 not found: ID does not exist" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.933949 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf9e0fb5-080e-4358-ae7c-de730482a384" (UID: "bf9e0fb5-080e-4358-ae7c-de730482a384"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.973777 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf9e0fb5-080e-4358-ae7c-de730482a384-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.997279 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:53 crc kubenswrapper[4751]: I1002 11:06:53.997937 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4cdnw"] Oct 02 11:06:55 crc kubenswrapper[4751]: I1002 11:06:55.558772 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" path="/var/lib/kubelet/pods/bf9e0fb5-080e-4358-ae7c-de730482a384/volumes" Oct 02 11:07:01 crc kubenswrapper[4751]: I1002 11:07:01.506764 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:07:01 crc kubenswrapper[4751]: I1002 11:07:01.507526 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216095 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216758 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216775 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216795 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="extract-utilities" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216803 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="extract-utilities" Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216817 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="extract-content" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216826 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="extract-content" Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216838 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="extract-content" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216845 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="extract-content" Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216857 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="extract-utilities" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216864 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="extract-utilities" Oct 02 11:07:08 crc kubenswrapper[4751]: E1002 11:07:08.216879 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216885 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.216990 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9e0fb5-080e-4358-ae7c-de730482a384" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.217005 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df81173-f48b-4f0c-9ea2-2f8bd7df0cb0" containerName="registry-server" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.217875 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.230652 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.365069 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.365474 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.365652 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k56hg\" (UniqueName: \"kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.466662 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.467074 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.467334 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.467382 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.467652 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k56hg\" (UniqueName: \"kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.490431 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k56hg\" (UniqueName: \"kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg\") pod \"redhat-marketplace-cd9w9\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.538053 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:08 crc kubenswrapper[4751]: I1002 11:07:08.754483 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:09 crc kubenswrapper[4751]: I1002 11:07:09.468097 4751 generic.go:334] "Generic (PLEG): container finished" podID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerID="a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe" exitCode=0 Oct 02 11:07:09 crc kubenswrapper[4751]: I1002 11:07:09.468406 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerDied","Data":"a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe"} Oct 02 11:07:09 crc kubenswrapper[4751]: I1002 11:07:09.468504 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerStarted","Data":"1cb761565f4a0696fed678f20608181c00548589f92520df17aa9406a195436f"} Oct 02 11:07:10 crc kubenswrapper[4751]: I1002 11:07:10.476406 4751 generic.go:334] "Generic (PLEG): container finished" podID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerID="f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1" exitCode=0 Oct 02 11:07:10 crc kubenswrapper[4751]: I1002 11:07:10.476508 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerDied","Data":"f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1"} Oct 02 11:07:11 crc kubenswrapper[4751]: I1002 11:07:11.486604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerStarted","Data":"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce"} Oct 02 11:07:11 crc kubenswrapper[4751]: I1002 11:07:11.509667 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cd9w9" podStartSLOduration=1.933039008 podStartE2EDuration="3.509638296s" podCreationTimestamp="2025-10-02 11:07:08 +0000 UTC" firstStartedPulling="2025-10-02 11:07:09.46999957 +0000 UTC m=+911.524226020" lastFinishedPulling="2025-10-02 11:07:11.046598858 +0000 UTC m=+913.100825308" observedRunningTime="2025-10-02 11:07:11.507582211 +0000 UTC m=+913.561808701" watchObservedRunningTime="2025-10-02 11:07:11.509638296 +0000 UTC m=+913.563864756" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.438367 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7"] Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.439930 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.442489 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.448473 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7"] Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.520415 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.520469 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2dtt\" (UniqueName: \"kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.520524 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.621574 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.621766 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.622139 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2dtt\" (UniqueName: \"kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.622469 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.622643 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.654599 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2dtt\" (UniqueName: \"kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.756228 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:12 crc kubenswrapper[4751]: I1002 11:07:12.973990 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7"] Oct 02 11:07:12 crc kubenswrapper[4751]: W1002 11:07:12.978414 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cdb897a_9c15_4d49_b6c1_8f6418b2edc7.slice/crio-de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b WatchSource:0}: Error finding container de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b: Status 404 returned error can't find the container with id de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b Oct 02 11:07:13 crc kubenswrapper[4751]: I1002 11:07:13.500635 4751 generic.go:334] "Generic (PLEG): container finished" podID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerID="534b2682cfe0491f4fc0fbbdb57a59a429e600b0d6be4f34b784761923f51cdc" exitCode=0 Oct 02 11:07:13 crc kubenswrapper[4751]: I1002 11:07:13.500688 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" event={"ID":"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7","Type":"ContainerDied","Data":"534b2682cfe0491f4fc0fbbdb57a59a429e600b0d6be4f34b784761923f51cdc"} Oct 02 11:07:13 crc kubenswrapper[4751]: I1002 11:07:13.500720 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" event={"ID":"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7","Type":"ContainerStarted","Data":"de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b"} Oct 02 11:07:15 crc kubenswrapper[4751]: I1002 11:07:15.514383 4751 generic.go:334] "Generic (PLEG): container finished" podID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerID="b88691a5d6a12cff3065718b09219cd2c144366dbcb172a2fa4839e5810b1447" exitCode=0 Oct 02 11:07:15 crc kubenswrapper[4751]: I1002 11:07:15.514457 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" event={"ID":"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7","Type":"ContainerDied","Data":"b88691a5d6a12cff3065718b09219cd2c144366dbcb172a2fa4839e5810b1447"} Oct 02 11:07:16 crc kubenswrapper[4751]: I1002 11:07:16.526204 4751 generic.go:334] "Generic (PLEG): container finished" podID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerID="fed9353486c660d5088f14337d64990ff13618d89b649b92252294d735f7eb5a" exitCode=0 Oct 02 11:07:16 crc kubenswrapper[4751]: I1002 11:07:16.526295 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" event={"ID":"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7","Type":"ContainerDied","Data":"fed9353486c660d5088f14337d64990ff13618d89b649b92252294d735f7eb5a"} Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.791495 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.989567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2dtt\" (UniqueName: \"kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt\") pod \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.989924 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util\") pod \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.990063 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle\") pod \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\" (UID: \"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7\") " Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.990844 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle" (OuterVolumeSpecName: "bundle") pod "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" (UID: "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:07:17 crc kubenswrapper[4751]: I1002 11:07:17.997907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt" (OuterVolumeSpecName: "kube-api-access-v2dtt") pod "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" (UID: "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7"). InnerVolumeSpecName "kube-api-access-v2dtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.009812 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util" (OuterVolumeSpecName: "util") pod "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" (UID: "5cdb897a-9c15-4d49-b6c1-8f6418b2edc7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.091021 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2dtt\" (UniqueName: \"kubernetes.io/projected/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-kube-api-access-v2dtt\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.091063 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-util\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.091073 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cdb897a-9c15-4d49-b6c1-8f6418b2edc7-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.538910 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.538962 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.541923 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" event={"ID":"5cdb897a-9c15-4d49-b6c1-8f6418b2edc7","Type":"ContainerDied","Data":"de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b"} Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.541952 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1b2846d42761b9ab4ae45db392c24d6a53b80e96b2d09304600175f896427b" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.542270 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7" Oct 02 11:07:18 crc kubenswrapper[4751]: I1002 11:07:18.580708 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:19 crc kubenswrapper[4751]: I1002 11:07:19.608323 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:19 crc kubenswrapper[4751]: I1002 11:07:19.654421 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:21 crc kubenswrapper[4751]: I1002 11:07:21.559284 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cd9w9" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="registry-server" containerID="cri-o://75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce" gracePeriod=2 Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.455853 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.567493 4751 generic.go:334] "Generic (PLEG): container finished" podID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerID="75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce" exitCode=0 Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.567546 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerDied","Data":"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce"} Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.567578 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cd9w9" event={"ID":"72b9c940-a97c-4abe-9aaf-89fa834eabb3","Type":"ContainerDied","Data":"1cb761565f4a0696fed678f20608181c00548589f92520df17aa9406a195436f"} Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.567598 4751 scope.go:117] "RemoveContainer" containerID="75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.567633 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cd9w9" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.588592 4751 scope.go:117] "RemoveContainer" containerID="f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.605391 4751 scope.go:117] "RemoveContainer" containerID="a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.621613 4751 scope.go:117] "RemoveContainer" containerID="75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce" Oct 02 11:07:22 crc kubenswrapper[4751]: E1002 11:07:22.622101 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce\": container with ID starting with 75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce not found: ID does not exist" containerID="75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.622162 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce"} err="failed to get container status \"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce\": rpc error: code = NotFound desc = could not find container \"75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce\": container with ID starting with 75a5b77868c1367d38ea97a8dbda51c1aeef4448b51bacada9ebca0aa5b4b4ce not found: ID does not exist" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.622238 4751 scope.go:117] "RemoveContainer" containerID="f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1" Oct 02 11:07:22 crc kubenswrapper[4751]: E1002 11:07:22.622702 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1\": container with ID starting with f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1 not found: ID does not exist" containerID="f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.622800 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1"} err="failed to get container status \"f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1\": rpc error: code = NotFound desc = could not find container \"f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1\": container with ID starting with f5dcf14a5a6b0a36152db39412b691d9fbab2421f8f8739d6256381237c001b1 not found: ID does not exist" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.622887 4751 scope.go:117] "RemoveContainer" containerID="a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe" Oct 02 11:07:22 crc kubenswrapper[4751]: E1002 11:07:22.623248 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe\": container with ID starting with a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe not found: ID does not exist" containerID="a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.623344 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe"} err="failed to get container status \"a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe\": rpc error: code = NotFound desc = could not find container \"a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe\": container with ID starting with a0d8096f0af470b71d0c440949455aea565d9e87db4f28784894d188d2a41abe not found: ID does not exist" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.642694 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities\") pod \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.642774 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content\") pod \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.642894 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k56hg\" (UniqueName: \"kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg\") pod \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\" (UID: \"72b9c940-a97c-4abe-9aaf-89fa834eabb3\") " Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.644118 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities" (OuterVolumeSpecName: "utilities") pod "72b9c940-a97c-4abe-9aaf-89fa834eabb3" (UID: "72b9c940-a97c-4abe-9aaf-89fa834eabb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.651081 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg" (OuterVolumeSpecName: "kube-api-access-k56hg") pod "72b9c940-a97c-4abe-9aaf-89fa834eabb3" (UID: "72b9c940-a97c-4abe-9aaf-89fa834eabb3"). InnerVolumeSpecName "kube-api-access-k56hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.657790 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72b9c940-a97c-4abe-9aaf-89fa834eabb3" (UID: "72b9c940-a97c-4abe-9aaf-89fa834eabb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.745024 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.745061 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k56hg\" (UniqueName: \"kubernetes.io/projected/72b9c940-a97c-4abe-9aaf-89fa834eabb3-kube-api-access-k56hg\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.745074 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b9c940-a97c-4abe-9aaf-89fa834eabb3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.899807 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:22 crc kubenswrapper[4751]: I1002 11:07:22.904526 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cd9w9"] Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404653 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p576x"] Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404880 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="extract" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404894 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="extract" Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404905 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="extract-utilities" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404912 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="extract-utilities" Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404919 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="util" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404926 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="util" Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404948 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="extract-content" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404957 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="extract-content" Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404969 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="registry-server" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404977 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="registry-server" Oct 02 11:07:23 crc kubenswrapper[4751]: E1002 11:07:23.404987 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="pull" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.404994 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="pull" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.405098 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" containerName="registry-server" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.405109 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cdb897a-9c15-4d49-b6c1-8f6418b2edc7" containerName="extract" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.405524 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.412305 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.412829 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.412887 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-5d2qk" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.429502 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p576x"] Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.553534 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm76b\" (UniqueName: \"kubernetes.io/projected/bf80f016-df98-4666-983e-d8537054d39d-kube-api-access-zm76b\") pod \"nmstate-operator-858ddd8f98-p576x\" (UID: \"bf80f016-df98-4666-983e-d8537054d39d\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.556953 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b9c940-a97c-4abe-9aaf-89fa834eabb3" path="/var/lib/kubelet/pods/72b9c940-a97c-4abe-9aaf-89fa834eabb3/volumes" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.655748 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm76b\" (UniqueName: \"kubernetes.io/projected/bf80f016-df98-4666-983e-d8537054d39d-kube-api-access-zm76b\") pod \"nmstate-operator-858ddd8f98-p576x\" (UID: \"bf80f016-df98-4666-983e-d8537054d39d\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.675219 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm76b\" (UniqueName: \"kubernetes.io/projected/bf80f016-df98-4666-983e-d8537054d39d-kube-api-access-zm76b\") pod \"nmstate-operator-858ddd8f98-p576x\" (UID: \"bf80f016-df98-4666-983e-d8537054d39d\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.718499 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" Oct 02 11:07:23 crc kubenswrapper[4751]: I1002 11:07:23.981724 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-p576x"] Oct 02 11:07:24 crc kubenswrapper[4751]: I1002 11:07:24.580807 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" event={"ID":"bf80f016-df98-4666-983e-d8537054d39d","Type":"ContainerStarted","Data":"1199ff97f9faf35ef6f038bf970b54bfb42a30fa845de8da0a107871d2236218"} Oct 02 11:07:26 crc kubenswrapper[4751]: I1002 11:07:26.592539 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" event={"ID":"bf80f016-df98-4666-983e-d8537054d39d","Type":"ContainerStarted","Data":"9dee08d1229cb9be0523cea7fd2d5b7d349797ed5511eb835a75f5977ec0089f"} Oct 02 11:07:26 crc kubenswrapper[4751]: I1002 11:07:26.614949 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-p576x" podStartSLOduration=1.499112116 podStartE2EDuration="3.614931521s" podCreationTimestamp="2025-10-02 11:07:23 +0000 UTC" firstStartedPulling="2025-10-02 11:07:23.991848366 +0000 UTC m=+926.046074816" lastFinishedPulling="2025-10-02 11:07:26.107667761 +0000 UTC m=+928.161894221" observedRunningTime="2025-10-02 11:07:26.611437676 +0000 UTC m=+928.665664126" watchObservedRunningTime="2025-10-02 11:07:26.614931521 +0000 UTC m=+928.669157971" Oct 02 11:07:31 crc kubenswrapper[4751]: I1002 11:07:31.507791 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:07:31 crc kubenswrapper[4751]: I1002 11:07:31.508221 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:07:33 crc kubenswrapper[4751]: I1002 11:07:33.985936 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww"] Oct 02 11:07:33 crc kubenswrapper[4751]: I1002 11:07:33.987289 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" Oct 02 11:07:33 crc kubenswrapper[4751]: I1002 11:07:33.989054 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-cbnvl" Oct 02 11:07:33 crc kubenswrapper[4751]: I1002 11:07:33.999024 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l"] Oct 02 11:07:33 crc kubenswrapper[4751]: I1002 11:07:33.999726 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.002732 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.006631 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-czmv9"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.007239 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.010162 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.025713 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.127000 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.127774 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.129196 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-m9wm9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.129529 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.129546 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.138472 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.185949 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsms6\" (UniqueName: \"kubernetes.io/projected/203e7187-9944-44f7-8a43-615567c3ab3c-kube-api-access-gsms6\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186015 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-dbus-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186128 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-ovs-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186195 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm5hz\" (UniqueName: \"kubernetes.io/projected/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-kube-api-access-sm5hz\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186218 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186233 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-nmstate-lock\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.186250 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj22p\" (UniqueName: \"kubernetes.io/projected/9b91b188-db18-4362-bf3e-612233c1aa7d-kube-api-access-mj22p\") pod \"nmstate-metrics-fdff9cb8d-6jmww\" (UID: \"9b91b188-db18-4362-bf3e-612233c1aa7d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287293 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rkzd\" (UniqueName: \"kubernetes.io/projected/fbc72901-fb4e-4476-b339-41c851edf1fe-kube-api-access-2rkzd\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287353 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbc72901-fb4e-4476-b339-41c851edf1fe-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287397 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm5hz\" (UniqueName: \"kubernetes.io/projected/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-kube-api-access-sm5hz\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287425 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj22p\" (UniqueName: \"kubernetes.io/projected/9b91b188-db18-4362-bf3e-612233c1aa7d-kube-api-access-mj22p\") pod \"nmstate-metrics-fdff9cb8d-6jmww\" (UID: \"9b91b188-db18-4362-bf3e-612233c1aa7d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287450 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287470 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-nmstate-lock\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287497 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsms6\" (UniqueName: \"kubernetes.io/projected/203e7187-9944-44f7-8a43-615567c3ab3c-kube-api-access-gsms6\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287551 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-dbus-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287587 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-ovs-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: E1002 11:07:34.287590 4751 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287618 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: E1002 11:07:34.287651 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair podName:b0a29038-b5ae-46b0-b5a2-4f7f952fec45 nodeName:}" failed. No retries permitted until 2025-10-02 11:07:34.787630223 +0000 UTC m=+936.841856673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair") pod "nmstate-webhook-6cdbc54649-9tq9l" (UID: "b0a29038-b5ae-46b0-b5a2-4f7f952fec45") : secret "openshift-nmstate-webhook" not found Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287695 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-nmstate-lock\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287771 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-ovs-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.287980 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/203e7187-9944-44f7-8a43-615567c3ab3c-dbus-socket\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.312872 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj22p\" (UniqueName: \"kubernetes.io/projected/9b91b188-db18-4362-bf3e-612233c1aa7d-kube-api-access-mj22p\") pod \"nmstate-metrics-fdff9cb8d-6jmww\" (UID: \"9b91b188-db18-4362-bf3e-612233c1aa7d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.312896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsms6\" (UniqueName: \"kubernetes.io/projected/203e7187-9944-44f7-8a43-615567c3ab3c-kube-api-access-gsms6\") pod \"nmstate-handler-czmv9\" (UID: \"203e7187-9944-44f7-8a43-615567c3ab3c\") " pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.316944 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm5hz\" (UniqueName: \"kubernetes.io/projected/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-kube-api-access-sm5hz\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.326582 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7b5885c8df-5ttd4"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.327449 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.340779 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.349723 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b5885c8df-5ttd4"] Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389012 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-service-ca\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-oauth-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389099 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-console-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389127 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389143 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-trusted-ca-bundle\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389162 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389191 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-oauth-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389213 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lccm\" (UniqueName: \"kubernetes.io/projected/914308fd-5784-4c59-939e-bd3dc51e7405-kube-api-access-6lccm\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389231 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rkzd\" (UniqueName: \"kubernetes.io/projected/fbc72901-fb4e-4476-b339-41c851edf1fe-kube-api-access-2rkzd\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.389250 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbc72901-fb4e-4476-b339-41c851edf1fe-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: E1002 11:07:34.389771 4751 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 02 11:07:34 crc kubenswrapper[4751]: E1002 11:07:34.389831 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert podName:fbc72901-fb4e-4476-b339-41c851edf1fe nodeName:}" failed. No retries permitted until 2025-10-02 11:07:34.889811839 +0000 UTC m=+936.944038289 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-dpwvm" (UID: "fbc72901-fb4e-4476-b339-41c851edf1fe") : secret "plugin-serving-cert" not found Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.390055 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbc72901-fb4e-4476-b339-41c851edf1fe-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.408041 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rkzd\" (UniqueName: \"kubernetes.io/projected/fbc72901-fb4e-4476-b339-41c851edf1fe-kube-api-access-2rkzd\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.489928 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-oauth-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490142 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-console-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490277 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-trusted-ca-bundle\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490355 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490434 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-oauth-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490504 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lccm\" (UniqueName: \"kubernetes.io/projected/914308fd-5784-4c59-939e-bd3dc51e7405-kube-api-access-6lccm\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.490607 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-service-ca\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.491511 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-service-ca\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.491913 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-console-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.492051 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-oauth-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.492160 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/914308fd-5784-4c59-939e-bd3dc51e7405-trusted-ca-bundle\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.493621 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-serving-cert\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.493755 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/914308fd-5784-4c59-939e-bd3dc51e7405-console-oauth-config\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.506560 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lccm\" (UniqueName: \"kubernetes.io/projected/914308fd-5784-4c59-939e-bd3dc51e7405-kube-api-access-6lccm\") pod \"console-7b5885c8df-5ttd4\" (UID: \"914308fd-5784-4c59-939e-bd3dc51e7405\") " pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.608693 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.634585 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-czmv9" event={"ID":"203e7187-9944-44f7-8a43-615567c3ab3c","Type":"ContainerStarted","Data":"3a48146ba71fc660b1662d726b8af37d335d59be8b1767f747710e36b8aaf392"} Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.701578 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.794447 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.799337 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0a29038-b5ae-46b0-b5a2-4f7f952fec45-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-9tq9l\" (UID: \"b0a29038-b5ae-46b0-b5a2-4f7f952fec45\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.842265 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww"] Oct 02 11:07:34 crc kubenswrapper[4751]: W1002 11:07:34.854127 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b91b188_db18_4362_bf3e_612233c1aa7d.slice/crio-73cd46d471ad52cb28668a5c87d58837a8fcea9d9d0e1543d17a3d85a80e03a2 WatchSource:0}: Error finding container 73cd46d471ad52cb28668a5c87d58837a8fcea9d9d0e1543d17a3d85a80e03a2: Status 404 returned error can't find the container with id 73cd46d471ad52cb28668a5c87d58837a8fcea9d9d0e1543d17a3d85a80e03a2 Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.896010 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.899655 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc72901-fb4e-4476-b339-41c851edf1fe-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-dpwvm\" (UID: \"fbc72901-fb4e-4476-b339-41c851edf1fe\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:34 crc kubenswrapper[4751]: I1002 11:07:34.926001 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.057502 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.106019 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l"] Oct 02 11:07:35 crc kubenswrapper[4751]: W1002 11:07:35.116658 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0a29038_b5ae_46b0_b5a2_4f7f952fec45.slice/crio-f2842777e97af57c2dee8ceb81868bb78cda446aea45dc3f501aeeadeaf68c92 WatchSource:0}: Error finding container f2842777e97af57c2dee8ceb81868bb78cda446aea45dc3f501aeeadeaf68c92: Status 404 returned error can't find the container with id f2842777e97af57c2dee8ceb81868bb78cda446aea45dc3f501aeeadeaf68c92 Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.122835 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b5885c8df-5ttd4"] Oct 02 11:07:35 crc kubenswrapper[4751]: W1002 11:07:35.140467 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod914308fd_5784_4c59_939e_bd3dc51e7405.slice/crio-925629707f72cd2c4d2f16a74a26d2bd004c0e3f2f79e9419fc0342e42c568c2 WatchSource:0}: Error finding container 925629707f72cd2c4d2f16a74a26d2bd004c0e3f2f79e9419fc0342e42c568c2: Status 404 returned error can't find the container with id 925629707f72cd2c4d2f16a74a26d2bd004c0e3f2f79e9419fc0342e42c568c2 Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.266515 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm"] Oct 02 11:07:35 crc kubenswrapper[4751]: W1002 11:07:35.272359 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbc72901_fb4e_4476_b339_41c851edf1fe.slice/crio-77d1997195b5df01cc7e260f00f77b6766a5c7176ceb73962b7269fab72512ed WatchSource:0}: Error finding container 77d1997195b5df01cc7e260f00f77b6766a5c7176ceb73962b7269fab72512ed: Status 404 returned error can't find the container with id 77d1997195b5df01cc7e260f00f77b6766a5c7176ceb73962b7269fab72512ed Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.641918 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" event={"ID":"b0a29038-b5ae-46b0-b5a2-4f7f952fec45","Type":"ContainerStarted","Data":"f2842777e97af57c2dee8ceb81868bb78cda446aea45dc3f501aeeadeaf68c92"} Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.644004 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b5885c8df-5ttd4" event={"ID":"914308fd-5784-4c59-939e-bd3dc51e7405","Type":"ContainerStarted","Data":"49d99853a09d0f3c8b51585b81c752e8498d835c438985217354522a7b50162b"} Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.644047 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b5885c8df-5ttd4" event={"ID":"914308fd-5784-4c59-939e-bd3dc51e7405","Type":"ContainerStarted","Data":"925629707f72cd2c4d2f16a74a26d2bd004c0e3f2f79e9419fc0342e42c568c2"} Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.646594 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" event={"ID":"9b91b188-db18-4362-bf3e-612233c1aa7d","Type":"ContainerStarted","Data":"73cd46d471ad52cb28668a5c87d58837a8fcea9d9d0e1543d17a3d85a80e03a2"} Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.647829 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" event={"ID":"fbc72901-fb4e-4476-b339-41c851edf1fe","Type":"ContainerStarted","Data":"77d1997195b5df01cc7e260f00f77b6766a5c7176ceb73962b7269fab72512ed"} Oct 02 11:07:35 crc kubenswrapper[4751]: I1002 11:07:35.663347 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7b5885c8df-5ttd4" podStartSLOduration=1.663326734 podStartE2EDuration="1.663326734s" podCreationTimestamp="2025-10-02 11:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:07:35.661271758 +0000 UTC m=+937.715498248" watchObservedRunningTime="2025-10-02 11:07:35.663326734 +0000 UTC m=+937.717553174" Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.675115 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-czmv9" event={"ID":"203e7187-9944-44f7-8a43-615567c3ab3c","Type":"ContainerStarted","Data":"02a3cdc43ae89a5e66d2dc7764eea0a6bde7b9ff0a3ede94ab79104ce28cb2b8"} Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.675747 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.678196 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" event={"ID":"9b91b188-db18-4362-bf3e-612233c1aa7d","Type":"ContainerStarted","Data":"5d97b9660142ffb16f9172adf0dc3e14cf1a560e5001ba4df3c9a8232f761774"} Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.680099 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" event={"ID":"fbc72901-fb4e-4476-b339-41c851edf1fe","Type":"ContainerStarted","Data":"bb8694339c0b4150c8d23e97d76524eb102a130b641255977f3ee9ab5e8aaf27"} Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.681838 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" event={"ID":"b0a29038-b5ae-46b0-b5a2-4f7f952fec45","Type":"ContainerStarted","Data":"e9b2bb5c8308e548a1ae078d953c911a08327c12ecd0cf9d59bd4e3a60db7ec2"} Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.682014 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.703315 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-czmv9" podStartSLOduration=2.133549252 podStartE2EDuration="5.703284452s" podCreationTimestamp="2025-10-02 11:07:33 +0000 UTC" firstStartedPulling="2025-10-02 11:07:34.368973143 +0000 UTC m=+936.423199593" lastFinishedPulling="2025-10-02 11:07:37.938708313 +0000 UTC m=+939.992934793" observedRunningTime="2025-10-02 11:07:38.698521743 +0000 UTC m=+940.752748233" watchObservedRunningTime="2025-10-02 11:07:38.703284452 +0000 UTC m=+940.757510942" Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.715950 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" podStartSLOduration=2.864479448 podStartE2EDuration="5.715924006s" podCreationTimestamp="2025-10-02 11:07:33 +0000 UTC" firstStartedPulling="2025-10-02 11:07:35.120612571 +0000 UTC m=+937.174839061" lastFinishedPulling="2025-10-02 11:07:37.972057129 +0000 UTC m=+940.026283619" observedRunningTime="2025-10-02 11:07:38.714357703 +0000 UTC m=+940.768584223" watchObservedRunningTime="2025-10-02 11:07:38.715924006 +0000 UTC m=+940.770150486" Oct 02 11:07:38 crc kubenswrapper[4751]: I1002 11:07:38.732612 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-dpwvm" podStartSLOduration=2.066777344 podStartE2EDuration="4.732565838s" podCreationTimestamp="2025-10-02 11:07:34 +0000 UTC" firstStartedPulling="2025-10-02 11:07:35.27812657 +0000 UTC m=+937.332353020" lastFinishedPulling="2025-10-02 11:07:37.943915064 +0000 UTC m=+939.998141514" observedRunningTime="2025-10-02 11:07:38.729823743 +0000 UTC m=+940.784050233" watchObservedRunningTime="2025-10-02 11:07:38.732565838 +0000 UTC m=+940.786792328" Oct 02 11:07:41 crc kubenswrapper[4751]: I1002 11:07:41.700221 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" event={"ID":"9b91b188-db18-4362-bf3e-612233c1aa7d","Type":"ContainerStarted","Data":"15475738ebb0cecdc75e7f3af334cfc193f1d7f53ad697ef033c0703d35df00c"} Oct 02 11:07:41 crc kubenswrapper[4751]: I1002 11:07:41.713267 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6jmww" podStartSLOduration=2.9631448970000003 podStartE2EDuration="8.713239146s" podCreationTimestamp="2025-10-02 11:07:33 +0000 UTC" firstStartedPulling="2025-10-02 11:07:34.856738903 +0000 UTC m=+936.910965353" lastFinishedPulling="2025-10-02 11:07:40.606833132 +0000 UTC m=+942.661059602" observedRunningTime="2025-10-02 11:07:41.712625549 +0000 UTC m=+943.766852039" watchObservedRunningTime="2025-10-02 11:07:41.713239146 +0000 UTC m=+943.767465606" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.364703 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-czmv9" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.702419 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.702486 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.708325 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.723984 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7b5885c8df-5ttd4" Oct 02 11:07:44 crc kubenswrapper[4751]: I1002 11:07:44.778337 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 11:07:54 crc kubenswrapper[4751]: I1002 11:07:54.931840 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-9tq9l" Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.507323 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.507892 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.507934 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.508546 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.508599 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c" gracePeriod=600 Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.831193 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c" exitCode=0 Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.831204 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c"} Oct 02 11:08:01 crc kubenswrapper[4751]: I1002 11:08:01.831596 4751 scope.go:117] "RemoveContainer" containerID="fec5716e3cd8527ac2a3432a2ec10a07b79423b3422c3252441f9f8fbbcebaed" Oct 02 11:08:02 crc kubenswrapper[4751]: I1002 11:08:02.839605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182"} Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.140351 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz"] Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.141792 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.143900 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.149632 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz"] Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.336361 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.336506 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tdwl\" (UniqueName: \"kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.336608 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.437768 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.437843 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tdwl\" (UniqueName: \"kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.437903 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.438709 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.438756 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.461343 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tdwl\" (UniqueName: \"kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.757778 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:09 crc kubenswrapper[4751]: I1002 11:08:09.836225 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-k2g7v" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" containerID="cri-o://f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38" gracePeriod=15 Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.150592 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k2g7v_b419cc91-b70e-4108-a04a-761f0ce61313/console/0.log" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.150653 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.181848 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz"] Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349309 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349394 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349418 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349490 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349544 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.349608 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xlnq\" (UniqueName: \"kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq\") pod \"b419cc91-b70e-4108-a04a-761f0ce61313\" (UID: \"b419cc91-b70e-4108-a04a-761f0ce61313\") " Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.350296 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.350346 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config" (OuterVolumeSpecName: "console-config") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.350547 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.350682 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca" (OuterVolumeSpecName: "service-ca") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.355335 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.355360 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq" (OuterVolumeSpecName: "kube-api-access-4xlnq") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "kube-api-access-4xlnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.355940 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b419cc91-b70e-4108-a04a-761f0ce61313" (UID: "b419cc91-b70e-4108-a04a-761f0ce61313"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451322 4751 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451366 4751 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451377 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xlnq\" (UniqueName: \"kubernetes.io/projected/b419cc91-b70e-4108-a04a-761f0ce61313-kube-api-access-4xlnq\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451389 4751 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451400 4751 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b419cc91-b70e-4108-a04a-761f0ce61313-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451412 4751 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.451423 4751 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b419cc91-b70e-4108-a04a-761f0ce61313-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883407 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-k2g7v_b419cc91-b70e-4108-a04a-761f0ce61313/console/0.log" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883469 4751 generic.go:334] "Generic (PLEG): container finished" podID="b419cc91-b70e-4108-a04a-761f0ce61313" containerID="f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38" exitCode=2 Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883555 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-k2g7v" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883575 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k2g7v" event={"ID":"b419cc91-b70e-4108-a04a-761f0ce61313","Type":"ContainerDied","Data":"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38"} Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883674 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-k2g7v" event={"ID":"b419cc91-b70e-4108-a04a-761f0ce61313","Type":"ContainerDied","Data":"6bcf9d119ff9dbfcf8ef5894d22fdc019bf78c86d48db199f5b6dae253825bed"} Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.883712 4751 scope.go:117] "RemoveContainer" containerID="f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.885788 4751 generic.go:334] "Generic (PLEG): container finished" podID="020abe52-0534-45cd-845e-5680d8359e6b" containerID="be21d3429afd782471d9db24378d240bb3d4dad0419518b84e3476e4f627bb11" exitCode=0 Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.885824 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" event={"ID":"020abe52-0534-45cd-845e-5680d8359e6b","Type":"ContainerDied","Data":"be21d3429afd782471d9db24378d240bb3d4dad0419518b84e3476e4f627bb11"} Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.885847 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" event={"ID":"020abe52-0534-45cd-845e-5680d8359e6b","Type":"ContainerStarted","Data":"fd6606e54f8a81cd2f1a8e2c955b7c6d539e0ab65073f351f2bfd898ee6d6d20"} Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.905351 4751 scope.go:117] "RemoveContainer" containerID="f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38" Oct 02 11:08:10 crc kubenswrapper[4751]: E1002 11:08:10.906228 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38\": container with ID starting with f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38 not found: ID does not exist" containerID="f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.906283 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38"} err="failed to get container status \"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38\": rpc error: code = NotFound desc = could not find container \"f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38\": container with ID starting with f914be5b789b460c89d0e9324800e894c63500c1ed5471bbbf384998aa116c38 not found: ID does not exist" Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.921342 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 11:08:10 crc kubenswrapper[4751]: I1002 11:08:10.923941 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-k2g7v"] Oct 02 11:08:11 crc kubenswrapper[4751]: I1002 11:08:11.568784 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" path="/var/lib/kubelet/pods/b419cc91-b70e-4108-a04a-761f0ce61313/volumes" Oct 02 11:08:12 crc kubenswrapper[4751]: I1002 11:08:12.903735 4751 generic.go:334] "Generic (PLEG): container finished" podID="020abe52-0534-45cd-845e-5680d8359e6b" containerID="648cc404d0963b51f281d88a24e3eb5557e38cbfc88aaeee40f4ad5fcbb82623" exitCode=0 Oct 02 11:08:12 crc kubenswrapper[4751]: I1002 11:08:12.903812 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" event={"ID":"020abe52-0534-45cd-845e-5680d8359e6b","Type":"ContainerDied","Data":"648cc404d0963b51f281d88a24e3eb5557e38cbfc88aaeee40f4ad5fcbb82623"} Oct 02 11:08:13 crc kubenswrapper[4751]: I1002 11:08:13.911932 4751 generic.go:334] "Generic (PLEG): container finished" podID="020abe52-0534-45cd-845e-5680d8359e6b" containerID="ffa9c15bf7d0325ed6458552f2e54e9afe3a164bf76e6ec9f6321a650a046db7" exitCode=0 Oct 02 11:08:13 crc kubenswrapper[4751]: I1002 11:08:13.911977 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" event={"ID":"020abe52-0534-45cd-845e-5680d8359e6b","Type":"ContainerDied","Data":"ffa9c15bf7d0325ed6458552f2e54e9afe3a164bf76e6ec9f6321a650a046db7"} Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.178854 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.312006 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util\") pod \"020abe52-0534-45cd-845e-5680d8359e6b\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.312128 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle\") pod \"020abe52-0534-45cd-845e-5680d8359e6b\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.312183 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tdwl\" (UniqueName: \"kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl\") pod \"020abe52-0534-45cd-845e-5680d8359e6b\" (UID: \"020abe52-0534-45cd-845e-5680d8359e6b\") " Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.313843 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle" (OuterVolumeSpecName: "bundle") pod "020abe52-0534-45cd-845e-5680d8359e6b" (UID: "020abe52-0534-45cd-845e-5680d8359e6b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.317923 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl" (OuterVolumeSpecName: "kube-api-access-4tdwl") pod "020abe52-0534-45cd-845e-5680d8359e6b" (UID: "020abe52-0534-45cd-845e-5680d8359e6b"). InnerVolumeSpecName "kube-api-access-4tdwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.336094 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util" (OuterVolumeSpecName: "util") pod "020abe52-0534-45cd-845e-5680d8359e6b" (UID: "020abe52-0534-45cd-845e-5680d8359e6b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.413395 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-util\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.413434 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/020abe52-0534-45cd-845e-5680d8359e6b-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.413448 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tdwl\" (UniqueName: \"kubernetes.io/projected/020abe52-0534-45cd-845e-5680d8359e6b-kube-api-access-4tdwl\") on node \"crc\" DevicePath \"\"" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.927109 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" event={"ID":"020abe52-0534-45cd-845e-5680d8359e6b","Type":"ContainerDied","Data":"fd6606e54f8a81cd2f1a8e2c955b7c6d539e0ab65073f351f2bfd898ee6d6d20"} Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.927243 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd6606e54f8a81cd2f1a8e2c955b7c6d539e0ab65073f351f2bfd898ee6d6d20" Oct 02 11:08:15 crc kubenswrapper[4751]: I1002 11:08:15.927145 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.916447 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54"] Oct 02 11:08:24 crc kubenswrapper[4751]: E1002 11:08:24.917137 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="util" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917149 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="util" Oct 02 11:08:24 crc kubenswrapper[4751]: E1002 11:08:24.917162 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="extract" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917183 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="extract" Oct 02 11:08:24 crc kubenswrapper[4751]: E1002 11:08:24.917195 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917200 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" Oct 02 11:08:24 crc kubenswrapper[4751]: E1002 11:08:24.917212 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="pull" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917218 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="pull" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917311 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="020abe52-0534-45cd-845e-5680d8359e6b" containerName="extract" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917322 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="b419cc91-b70e-4108-a04a-761f0ce61313" containerName="console" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.917703 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.921365 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-npj75" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.921455 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.921385 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.921570 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.922204 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 11:08:24 crc kubenswrapper[4751]: I1002 11:08:24.937944 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54"] Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.028735 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2z6v\" (UniqueName: \"kubernetes.io/projected/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-kube-api-access-f2z6v\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.029185 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-webhook-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.029211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-apiservice-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.130496 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-webhook-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.130930 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-apiservice-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.131067 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2z6v\" (UniqueName: \"kubernetes.io/projected/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-kube-api-access-f2z6v\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.137214 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-webhook-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.148081 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2z6v\" (UniqueName: \"kubernetes.io/projected/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-kube-api-access-f2z6v\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.148125 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b846a35-2b40-4cfd-894c-94d9a3bd1a8e-apiservice-cert\") pod \"metallb-operator-controller-manager-66b7df5dd4-ghn54\" (UID: \"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e\") " pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.176325 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp"] Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.177106 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.178696 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.179201 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-z49r6" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.179909 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.197419 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp"] Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.234949 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.332781 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbwhs\" (UniqueName: \"kubernetes.io/projected/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-kube-api-access-dbwhs\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.333053 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-apiservice-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.333084 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-webhook-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.426026 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54"] Oct 02 11:08:25 crc kubenswrapper[4751]: W1002 11:08:25.433532 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b846a35_2b40_4cfd_894c_94d9a3bd1a8e.slice/crio-b489b2472c096bc698ad662f3f17593d12c462b4d14912de0ebf05c9bc14a799 WatchSource:0}: Error finding container b489b2472c096bc698ad662f3f17593d12c462b4d14912de0ebf05c9bc14a799: Status 404 returned error can't find the container with id b489b2472c096bc698ad662f3f17593d12c462b4d14912de0ebf05c9bc14a799 Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.433939 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbwhs\" (UniqueName: \"kubernetes.io/projected/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-kube-api-access-dbwhs\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.433973 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-apiservice-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.434014 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-webhook-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.439123 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-apiservice-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.439460 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-webhook-cert\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.450216 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbwhs\" (UniqueName: \"kubernetes.io/projected/4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece-kube-api-access-dbwhs\") pod \"metallb-operator-webhook-server-6c5dfc98c6-msmjp\" (UID: \"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece\") " pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.493568 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.709140 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp"] Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.982356 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" event={"ID":"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e","Type":"ContainerStarted","Data":"b489b2472c096bc698ad662f3f17593d12c462b4d14912de0ebf05c9bc14a799"} Oct 02 11:08:25 crc kubenswrapper[4751]: I1002 11:08:25.983478 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" event={"ID":"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece","Type":"ContainerStarted","Data":"06cc129b808b9a463698b0cbb0c4bc2720b6edfdcc58c5658647dbf3ea64ccd0"} Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.012632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" event={"ID":"4b846a35-2b40-4cfd-894c-94d9a3bd1a8e","Type":"ContainerStarted","Data":"d2b6c7a7599b820b200bbf5d198a61baf71aa4eeeb4bd1cb84daff4baf976e6a"} Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.013158 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.013892 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" event={"ID":"4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece","Type":"ContainerStarted","Data":"249d2565b9c06dc70a849993a184bde840f8d1a1b79b30ad516c35252f1b2f6e"} Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.014120 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.031690 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" podStartSLOduration=2.136696943 podStartE2EDuration="7.031666042s" podCreationTimestamp="2025-10-02 11:08:24 +0000 UTC" firstStartedPulling="2025-10-02 11:08:25.437079358 +0000 UTC m=+987.491305808" lastFinishedPulling="2025-10-02 11:08:30.332048457 +0000 UTC m=+992.386274907" observedRunningTime="2025-10-02 11:08:31.029183205 +0000 UTC m=+993.083409655" watchObservedRunningTime="2025-10-02 11:08:31.031666042 +0000 UTC m=+993.085892502" Oct 02 11:08:31 crc kubenswrapper[4751]: I1002 11:08:31.050726 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" podStartSLOduration=1.42303357 podStartE2EDuration="6.050705139s" podCreationTimestamp="2025-10-02 11:08:25 +0000 UTC" firstStartedPulling="2025-10-02 11:08:25.71915743 +0000 UTC m=+987.773383880" lastFinishedPulling="2025-10-02 11:08:30.346828999 +0000 UTC m=+992.401055449" observedRunningTime="2025-10-02 11:08:31.046115005 +0000 UTC m=+993.100341475" watchObservedRunningTime="2025-10-02 11:08:31.050705139 +0000 UTC m=+993.104931589" Oct 02 11:08:45 crc kubenswrapper[4751]: I1002 11:08:45.499045 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6c5dfc98c6-msmjp" Oct 02 11:09:05 crc kubenswrapper[4751]: I1002 11:09:05.239061 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66b7df5dd4-ghn54" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.053634 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.054771 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.056518 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-b9fs5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.056755 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.059485 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-cq9wn"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.063414 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.067773 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.068693 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.072966 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.131453 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tpkx9"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.132281 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.134421 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.134526 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.134808 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.134885 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-m6fxv" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.144008 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-2dzwd"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.144838 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.146532 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155623 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155667 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-sockets\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155687 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqx7h\" (UniqueName: \"kubernetes.io/projected/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-kube-api-access-qqx7h\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155762 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-startup\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155789 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssk7j\" (UniqueName: \"kubernetes.io/projected/c82c3ef0-d1d6-45d7-902b-100cfd482853-kube-api-access-ssk7j\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155815 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155863 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-conf\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155883 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-reloader\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.155900 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics-certs\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.170922 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2dzwd"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257161 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-startup\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257235 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-cert\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssk7j\" (UniqueName: \"kubernetes.io/projected/c82c3ef0-d1d6-45d7-902b-100cfd482853-kube-api-access-ssk7j\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257308 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-metrics-certs\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257370 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4nrt\" (UniqueName: \"kubernetes.io/projected/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-kube-api-access-m4nrt\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257403 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt98w\" (UniqueName: \"kubernetes.io/projected/46040bd9-f688-4966-a105-8ad5f93a1c8a-kube-api-access-jt98w\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257453 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-conf\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257475 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/46040bd9-f688-4966-a105-8ad5f93a1c8a-metallb-excludel2\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257501 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-reloader\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257525 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257549 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics-certs\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257576 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257607 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-sockets\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257633 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqx7h\" (UniqueName: \"kubernetes.io/projected/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-kube-api-access-qqx7h\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.257960 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-conf\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.258121 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-startup\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.258312 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.258389 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-reloader\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.258460 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c82c3ef0-d1d6-45d7-902b-100cfd482853-frr-sockets\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.265954 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c82c3ef0-d1d6-45d7-902b-100cfd482853-metrics-certs\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.270371 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.274247 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssk7j\" (UniqueName: \"kubernetes.io/projected/c82c3ef0-d1d6-45d7-902b-100cfd482853-kube-api-access-ssk7j\") pod \"frr-k8s-cq9wn\" (UID: \"c82c3ef0-d1d6-45d7-902b-100cfd482853\") " pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.276713 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqx7h\" (UniqueName: \"kubernetes.io/projected/6005ba7f-c783-4bc6-8df6-d045f6a9d80b-kube-api-access-qqx7h\") pod \"frr-k8s-webhook-server-64bf5d555-pvfc5\" (UID: \"6005ba7f-c783-4bc6-8df6-d045f6a9d80b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.359950 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4nrt\" (UniqueName: \"kubernetes.io/projected/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-kube-api-access-m4nrt\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360000 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360023 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt98w\" (UniqueName: \"kubernetes.io/projected/46040bd9-f688-4966-a105-8ad5f93a1c8a-kube-api-access-jt98w\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360046 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/46040bd9-f688-4966-a105-8ad5f93a1c8a-metallb-excludel2\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360071 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360120 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-cert\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.360141 4751 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.360238 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist podName:46040bd9-f688-4966-a105-8ad5f93a1c8a nodeName:}" failed. No retries permitted until 2025-10-02 11:09:06.860219591 +0000 UTC m=+1028.914446041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist") pod "speaker-tpkx9" (UID: "46040bd9-f688-4966-a105-8ad5f93a1c8a") : secret "metallb-memberlist" not found Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.360155 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-metrics-certs\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.360824 4751 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.360886 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs podName:46040bd9-f688-4966-a105-8ad5f93a1c8a nodeName:}" failed. No retries permitted until 2025-10-02 11:09:06.860865989 +0000 UTC m=+1028.915092439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs") pod "speaker-tpkx9" (UID: "46040bd9-f688-4966-a105-8ad5f93a1c8a") : secret "speaker-certs-secret" not found Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.361327 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/46040bd9-f688-4966-a105-8ad5f93a1c8a-metallb-excludel2\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.362882 4751 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.363716 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-metrics-certs\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.374232 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-cert\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.375669 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4nrt\" (UniqueName: \"kubernetes.io/projected/82465ad7-d9ca-4475-a0dc-6ce4068f20d2-kube-api-access-m4nrt\") pod \"controller-68d546b9d8-2dzwd\" (UID: \"82465ad7-d9ca-4475-a0dc-6ce4068f20d2\") " pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.382780 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt98w\" (UniqueName: \"kubernetes.io/projected/46040bd9-f688-4966-a105-8ad5f93a1c8a-kube-api-access-jt98w\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.392033 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.396603 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.461753 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.572976 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5"] Oct 02 11:09:06 crc kubenswrapper[4751]: W1002 11:09:06.581047 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6005ba7f_c783_4bc6_8df6_d045f6a9d80b.slice/crio-0a4b1c234a9f1cd7813248840142f4b65e302edf67cc28644a373a3475c73a29 WatchSource:0}: Error finding container 0a4b1c234a9f1cd7813248840142f4b65e302edf67cc28644a373a3475c73a29: Status 404 returned error can't find the container with id 0a4b1c234a9f1cd7813248840142f4b65e302edf67cc28644a373a3475c73a29 Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.652728 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2dzwd"] Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.874597 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.875070 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.874816 4751 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 11:09:06 crc kubenswrapper[4751]: E1002 11:09:06.875223 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist podName:46040bd9-f688-4966-a105-8ad5f93a1c8a nodeName:}" failed. No retries permitted until 2025-10-02 11:09:07.875196931 +0000 UTC m=+1029.929423401 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist") pod "speaker-tpkx9" (UID: "46040bd9-f688-4966-a105-8ad5f93a1c8a") : secret "metallb-memberlist" not found Oct 02 11:09:06 crc kubenswrapper[4751]: I1002 11:09:06.881461 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-metrics-certs\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.250451 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" event={"ID":"6005ba7f-c783-4bc6-8df6-d045f6a9d80b","Type":"ContainerStarted","Data":"0a4b1c234a9f1cd7813248840142f4b65e302edf67cc28644a373a3475c73a29"} Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.254993 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2dzwd" event={"ID":"82465ad7-d9ca-4475-a0dc-6ce4068f20d2","Type":"ContainerStarted","Data":"d0a7b521508d335f605cdc8a15b66a123160fa847e2ac5587030fdf3b56537e2"} Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.255121 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2dzwd" event={"ID":"82465ad7-d9ca-4475-a0dc-6ce4068f20d2","Type":"ContainerStarted","Data":"a1d55b78e68c3e4a762c2f054f8da5f1afb68c20847e6338bdcea73591fb52ca"} Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.255219 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.255292 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2dzwd" event={"ID":"82465ad7-d9ca-4475-a0dc-6ce4068f20d2","Type":"ContainerStarted","Data":"fdfa807009b69bfa9e9fdf4c273a0bd788c0db593802cac8c74460866153a2fb"} Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.256696 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"8d413b8ee47b6d484021d252fcbcc5afdab4ee4e20181f47188414bde326b4c5"} Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.277860 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-2dzwd" podStartSLOduration=1.277841907 podStartE2EDuration="1.277841907s" podCreationTimestamp="2025-10-02 11:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:09:07.275697039 +0000 UTC m=+1029.329923519" watchObservedRunningTime="2025-10-02 11:09:07.277841907 +0000 UTC m=+1029.332068357" Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.898879 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.903228 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/46040bd9-f688-4966-a105-8ad5f93a1c8a-memberlist\") pod \"speaker-tpkx9\" (UID: \"46040bd9-f688-4966-a105-8ad5f93a1c8a\") " pod="metallb-system/speaker-tpkx9" Oct 02 11:09:07 crc kubenswrapper[4751]: I1002 11:09:07.949265 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tpkx9" Oct 02 11:09:07 crc kubenswrapper[4751]: W1002 11:09:07.984661 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46040bd9_f688_4966_a105_8ad5f93a1c8a.slice/crio-1f7ea820016f7d273601fd5782ba666973180e9fbdbdf936ba7381ae536f4722 WatchSource:0}: Error finding container 1f7ea820016f7d273601fd5782ba666973180e9fbdbdf936ba7381ae536f4722: Status 404 returned error can't find the container with id 1f7ea820016f7d273601fd5782ba666973180e9fbdbdf936ba7381ae536f4722 Oct 02 11:09:08 crc kubenswrapper[4751]: I1002 11:09:08.265000 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tpkx9" event={"ID":"46040bd9-f688-4966-a105-8ad5f93a1c8a","Type":"ContainerStarted","Data":"826f63c240eaca67cfdc2e68078c87be2a7d785c6b02c59f644f56a23be75045"} Oct 02 11:09:08 crc kubenswrapper[4751]: I1002 11:09:08.265054 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tpkx9" event={"ID":"46040bd9-f688-4966-a105-8ad5f93a1c8a","Type":"ContainerStarted","Data":"1f7ea820016f7d273601fd5782ba666973180e9fbdbdf936ba7381ae536f4722"} Oct 02 11:09:09 crc kubenswrapper[4751]: I1002 11:09:09.279182 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tpkx9" event={"ID":"46040bd9-f688-4966-a105-8ad5f93a1c8a","Type":"ContainerStarted","Data":"42020772ad180348240c69590354534402012bf2fc116bd2abb1a164736bc48f"} Oct 02 11:09:09 crc kubenswrapper[4751]: I1002 11:09:09.281873 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tpkx9" Oct 02 11:09:09 crc kubenswrapper[4751]: I1002 11:09:09.306462 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tpkx9" podStartSLOduration=3.306448134 podStartE2EDuration="3.306448134s" podCreationTimestamp="2025-10-02 11:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:09:09.303048001 +0000 UTC m=+1031.357274451" watchObservedRunningTime="2025-10-02 11:09:09.306448134 +0000 UTC m=+1031.360674574" Oct 02 11:09:14 crc kubenswrapper[4751]: I1002 11:09:14.319585 4751 generic.go:334] "Generic (PLEG): container finished" podID="c82c3ef0-d1d6-45d7-902b-100cfd482853" containerID="7fd7398b2f4d4b58466f5f1781799c94102793d44575a2af64e9ad145ced1707" exitCode=0 Oct 02 11:09:14 crc kubenswrapper[4751]: I1002 11:09:14.319674 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerDied","Data":"7fd7398b2f4d4b58466f5f1781799c94102793d44575a2af64e9ad145ced1707"} Oct 02 11:09:14 crc kubenswrapper[4751]: I1002 11:09:14.322686 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" event={"ID":"6005ba7f-c783-4bc6-8df6-d045f6a9d80b","Type":"ContainerStarted","Data":"7ee91b0c7a77b4f9c55ed45fedc11b4bd00959b3b8d8a1b9a278efcd5303b6c5"} Oct 02 11:09:14 crc kubenswrapper[4751]: I1002 11:09:14.323086 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:15 crc kubenswrapper[4751]: I1002 11:09:15.331408 4751 generic.go:334] "Generic (PLEG): container finished" podID="c82c3ef0-d1d6-45d7-902b-100cfd482853" containerID="3de42cbaf6f2801b9ed5443ec1ba26ded1966f7338e94c776c87d181ce98cdd2" exitCode=0 Oct 02 11:09:15 crc kubenswrapper[4751]: I1002 11:09:15.331453 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerDied","Data":"3de42cbaf6f2801b9ed5443ec1ba26ded1966f7338e94c776c87d181ce98cdd2"} Oct 02 11:09:15 crc kubenswrapper[4751]: I1002 11:09:15.353942 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" podStartSLOduration=2.434142947 podStartE2EDuration="9.353921669s" podCreationTimestamp="2025-10-02 11:09:06 +0000 UTC" firstStartedPulling="2025-10-02 11:09:06.582958452 +0000 UTC m=+1028.637184902" lastFinishedPulling="2025-10-02 11:09:13.502737174 +0000 UTC m=+1035.556963624" observedRunningTime="2025-10-02 11:09:14.382211743 +0000 UTC m=+1036.436438253" watchObservedRunningTime="2025-10-02 11:09:15.353921669 +0000 UTC m=+1037.408148119" Oct 02 11:09:16 crc kubenswrapper[4751]: I1002 11:09:16.339946 4751 generic.go:334] "Generic (PLEG): container finished" podID="c82c3ef0-d1d6-45d7-902b-100cfd482853" containerID="f29cf6084370733996f5efcadc2b52e8e24ce3c5d5b9de4e949050061a5bee85" exitCode=0 Oct 02 11:09:16 crc kubenswrapper[4751]: I1002 11:09:16.339987 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerDied","Data":"f29cf6084370733996f5efcadc2b52e8e24ce3c5d5b9de4e949050061a5bee85"} Oct 02 11:09:16 crc kubenswrapper[4751]: I1002 11:09:16.466916 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-2dzwd" Oct 02 11:09:17 crc kubenswrapper[4751]: I1002 11:09:17.358524 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"734dab06f4ed7f93922da3598045d105cc65f60fca8fb89c58453c46e9cd52cd"} Oct 02 11:09:17 crc kubenswrapper[4751]: I1002 11:09:17.358595 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"e78f0df9783b11385328b42489548b722535ab62d8b34857ff1c33fa8f73c4d2"} Oct 02 11:09:17 crc kubenswrapper[4751]: I1002 11:09:17.358612 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"b7d0a62b774add91a03c04967e23e7a60a2152decf381ba69359924e98afd4ab"} Oct 02 11:09:17 crc kubenswrapper[4751]: I1002 11:09:17.358622 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"16f09c098e55ed21d448d08beb0d8d54199591b4cab7a06d1a8fa1c7ba0ff6b5"} Oct 02 11:09:17 crc kubenswrapper[4751]: I1002 11:09:17.358632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"edb44b1689cd2f7a25bfe665bd6e8e12bed3bc13f5de60cadcea902a8174fa7a"} Oct 02 11:09:18 crc kubenswrapper[4751]: I1002 11:09:18.368463 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-cq9wn" event={"ID":"c82c3ef0-d1d6-45d7-902b-100cfd482853","Type":"ContainerStarted","Data":"b1564563635b9d114465fc36b2bbc8d8af1a5a87ffc833ca8cb58c7275aaaa8f"} Oct 02 11:09:18 crc kubenswrapper[4751]: I1002 11:09:18.368784 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:18 crc kubenswrapper[4751]: I1002 11:09:18.395091 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-cq9wn" podStartSLOduration=5.364668653 podStartE2EDuration="12.39507227s" podCreationTimestamp="2025-10-02 11:09:06 +0000 UTC" firstStartedPulling="2025-10-02 11:09:06.50115831 +0000 UTC m=+1028.555384760" lastFinishedPulling="2025-10-02 11:09:13.531561887 +0000 UTC m=+1035.585788377" observedRunningTime="2025-10-02 11:09:18.391289328 +0000 UTC m=+1040.445515788" watchObservedRunningTime="2025-10-02 11:09:18.39507227 +0000 UTC m=+1040.449298740" Oct 02 11:09:21 crc kubenswrapper[4751]: I1002 11:09:21.393445 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:21 crc kubenswrapper[4751]: I1002 11:09:21.437587 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:26 crc kubenswrapper[4751]: I1002 11:09:26.395993 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-cq9wn" Oct 02 11:09:26 crc kubenswrapper[4751]: I1002 11:09:26.401402 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pvfc5" Oct 02 11:09:27 crc kubenswrapper[4751]: I1002 11:09:27.954794 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tpkx9" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.155965 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xpjtm"] Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.157341 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.160397 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.160788 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lf2w5" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.160795 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.168019 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xpjtm"] Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.251078 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2k9t\" (UniqueName: \"kubernetes.io/projected/374515d5-bfd6-4830-bfd8-e5511e512187-kube-api-access-x2k9t\") pod \"openstack-operator-index-xpjtm\" (UID: \"374515d5-bfd6-4830-bfd8-e5511e512187\") " pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.353116 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2k9t\" (UniqueName: \"kubernetes.io/projected/374515d5-bfd6-4830-bfd8-e5511e512187-kube-api-access-x2k9t\") pod \"openstack-operator-index-xpjtm\" (UID: \"374515d5-bfd6-4830-bfd8-e5511e512187\") " pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.379021 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2k9t\" (UniqueName: \"kubernetes.io/projected/374515d5-bfd6-4830-bfd8-e5511e512187-kube-api-access-x2k9t\") pod \"openstack-operator-index-xpjtm\" (UID: \"374515d5-bfd6-4830-bfd8-e5511e512187\") " pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.488669 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:41 crc kubenswrapper[4751]: I1002 11:09:41.777367 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xpjtm"] Oct 02 11:09:41 crc kubenswrapper[4751]: W1002 11:09:41.785757 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod374515d5_bfd6_4830_bfd8_e5511e512187.slice/crio-52c26896a7eb4ab805084f7e65b54e215d9e8f62cc00d79fe3e1cbc56d3f20cf WatchSource:0}: Error finding container 52c26896a7eb4ab805084f7e65b54e215d9e8f62cc00d79fe3e1cbc56d3f20cf: Status 404 returned error can't find the container with id 52c26896a7eb4ab805084f7e65b54e215d9e8f62cc00d79fe3e1cbc56d3f20cf Oct 02 11:09:42 crc kubenswrapper[4751]: I1002 11:09:42.534612 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xpjtm" event={"ID":"374515d5-bfd6-4830-bfd8-e5511e512187","Type":"ContainerStarted","Data":"52c26896a7eb4ab805084f7e65b54e215d9e8f62cc00d79fe3e1cbc56d3f20cf"} Oct 02 11:09:45 crc kubenswrapper[4751]: I1002 11:09:45.562908 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xpjtm" event={"ID":"374515d5-bfd6-4830-bfd8-e5511e512187","Type":"ContainerStarted","Data":"fff194a176635bf838e422f7e2058198a8e84cd751daac4426ddf86f79db378e"} Oct 02 11:09:45 crc kubenswrapper[4751]: I1002 11:09:45.579750 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xpjtm" podStartSLOduration=1.395043554 podStartE2EDuration="4.579726524s" podCreationTimestamp="2025-10-02 11:09:41 +0000 UTC" firstStartedPulling="2025-10-02 11:09:41.788702193 +0000 UTC m=+1063.842928643" lastFinishedPulling="2025-10-02 11:09:44.973385133 +0000 UTC m=+1067.027611613" observedRunningTime="2025-10-02 11:09:45.575873139 +0000 UTC m=+1067.630099609" watchObservedRunningTime="2025-10-02 11:09:45.579726524 +0000 UTC m=+1067.633952984" Oct 02 11:09:51 crc kubenswrapper[4751]: I1002 11:09:51.489627 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:51 crc kubenswrapper[4751]: I1002 11:09:51.490059 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:51 crc kubenswrapper[4751]: I1002 11:09:51.526868 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:51 crc kubenswrapper[4751]: I1002 11:09:51.622526 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xpjtm" Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.794501 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk"] Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.796185 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.799009 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-f6z27" Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.806595 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk"] Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.918031 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.918111 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghf7z\" (UniqueName: \"kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:52 crc kubenswrapper[4751]: I1002 11:09:52.918158 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.019771 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.019853 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghf7z\" (UniqueName: \"kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.019918 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.020386 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.020749 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.043880 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghf7z\" (UniqueName: \"kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z\") pod \"157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.115901 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.517821 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk"] Oct 02 11:09:53 crc kubenswrapper[4751]: I1002 11:09:53.604665 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" event={"ID":"b22b717b-33f6-403e-9541-2204576e92ae","Type":"ContainerStarted","Data":"58242f4d41087bf37f5bd8413aad4ee4a84b17396186f57e92c1c53b7270027e"} Oct 02 11:09:54 crc kubenswrapper[4751]: I1002 11:09:54.614094 4751 generic.go:334] "Generic (PLEG): container finished" podID="b22b717b-33f6-403e-9541-2204576e92ae" containerID="c55243ecfd0caf24fe0722b7c112b6fb50a4ddc76adb3d8d80781faca658ed74" exitCode=0 Oct 02 11:09:54 crc kubenswrapper[4751]: I1002 11:09:54.614476 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" event={"ID":"b22b717b-33f6-403e-9541-2204576e92ae","Type":"ContainerDied","Data":"c55243ecfd0caf24fe0722b7c112b6fb50a4ddc76adb3d8d80781faca658ed74"} Oct 02 11:09:55 crc kubenswrapper[4751]: I1002 11:09:55.624321 4751 generic.go:334] "Generic (PLEG): container finished" podID="b22b717b-33f6-403e-9541-2204576e92ae" containerID="a0a909e9d494fc141b4b7a3cb94e6bf30c9b570b635ab6e5cccfe2dbf7085e2d" exitCode=0 Oct 02 11:09:55 crc kubenswrapper[4751]: I1002 11:09:55.624368 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" event={"ID":"b22b717b-33f6-403e-9541-2204576e92ae","Type":"ContainerDied","Data":"a0a909e9d494fc141b4b7a3cb94e6bf30c9b570b635ab6e5cccfe2dbf7085e2d"} Oct 02 11:09:56 crc kubenswrapper[4751]: I1002 11:09:56.638459 4751 generic.go:334] "Generic (PLEG): container finished" podID="b22b717b-33f6-403e-9541-2204576e92ae" containerID="db5f1877e3b5848c1a97ec4223607993ee9ead3ba7fc7710edbeba4755a49ce5" exitCode=0 Oct 02 11:09:56 crc kubenswrapper[4751]: I1002 11:09:56.638536 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" event={"ID":"b22b717b-33f6-403e-9541-2204576e92ae","Type":"ContainerDied","Data":"db5f1877e3b5848c1a97ec4223607993ee9ead3ba7fc7710edbeba4755a49ce5"} Oct 02 11:09:57 crc kubenswrapper[4751]: I1002 11:09:57.895205 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.087744 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle\") pod \"b22b717b-33f6-403e-9541-2204576e92ae\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.087798 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghf7z\" (UniqueName: \"kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z\") pod \"b22b717b-33f6-403e-9541-2204576e92ae\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.087946 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util\") pod \"b22b717b-33f6-403e-9541-2204576e92ae\" (UID: \"b22b717b-33f6-403e-9541-2204576e92ae\") " Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.088693 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle" (OuterVolumeSpecName: "bundle") pod "b22b717b-33f6-403e-9541-2204576e92ae" (UID: "b22b717b-33f6-403e-9541-2204576e92ae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.095662 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z" (OuterVolumeSpecName: "kube-api-access-ghf7z") pod "b22b717b-33f6-403e-9541-2204576e92ae" (UID: "b22b717b-33f6-403e-9541-2204576e92ae"). InnerVolumeSpecName "kube-api-access-ghf7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.100903 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util" (OuterVolumeSpecName: "util") pod "b22b717b-33f6-403e-9541-2204576e92ae" (UID: "b22b717b-33f6-403e-9541-2204576e92ae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.190235 4751 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-util\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.190309 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghf7z\" (UniqueName: \"kubernetes.io/projected/b22b717b-33f6-403e-9541-2204576e92ae-kube-api-access-ghf7z\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.190333 4751 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b22b717b-33f6-403e-9541-2204576e92ae-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.657212 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" event={"ID":"b22b717b-33f6-403e-9541-2204576e92ae","Type":"ContainerDied","Data":"58242f4d41087bf37f5bd8413aad4ee4a84b17396186f57e92c1c53b7270027e"} Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.657261 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58242f4d41087bf37f5bd8413aad4ee4a84b17396186f57e92c1c53b7270027e" Oct 02 11:09:58 crc kubenswrapper[4751]: I1002 11:09:58.657272 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.462971 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526"] Oct 02 11:10:03 crc kubenswrapper[4751]: E1002 11:10:03.463597 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="pull" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.463612 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="pull" Oct 02 11:10:03 crc kubenswrapper[4751]: E1002 11:10:03.463622 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="extract" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.463628 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="extract" Oct 02 11:10:03 crc kubenswrapper[4751]: E1002 11:10:03.463645 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="util" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.463652 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="util" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.463771 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="b22b717b-33f6-403e-9541-2204576e92ae" containerName="extract" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.464417 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.469201 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qz8tz" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.490832 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526"] Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.658580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwwn4\" (UniqueName: \"kubernetes.io/projected/fa36b949-67d9-4930-a31a-0381cd077c4a-kube-api-access-mwwn4\") pod \"openstack-operator-controller-operator-cc764bd77-ws526\" (UID: \"fa36b949-67d9-4930-a31a-0381cd077c4a\") " pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.759593 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwwn4\" (UniqueName: \"kubernetes.io/projected/fa36b949-67d9-4930-a31a-0381cd077c4a-kube-api-access-mwwn4\") pod \"openstack-operator-controller-operator-cc764bd77-ws526\" (UID: \"fa36b949-67d9-4930-a31a-0381cd077c4a\") " pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.778524 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwwn4\" (UniqueName: \"kubernetes.io/projected/fa36b949-67d9-4930-a31a-0381cd077c4a-kube-api-access-mwwn4\") pod \"openstack-operator-controller-operator-cc764bd77-ws526\" (UID: \"fa36b949-67d9-4930-a31a-0381cd077c4a\") " pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:03 crc kubenswrapper[4751]: I1002 11:10:03.782747 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:04 crc kubenswrapper[4751]: I1002 11:10:04.203749 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526"] Oct 02 11:10:04 crc kubenswrapper[4751]: I1002 11:10:04.694351 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" event={"ID":"fa36b949-67d9-4930-a31a-0381cd077c4a","Type":"ContainerStarted","Data":"ea75b656b17bca49c77622875ae160a927a7f400575b33c17f29e6d23047be93"} Oct 02 11:10:08 crc kubenswrapper[4751]: I1002 11:10:08.723489 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" event={"ID":"fa36b949-67d9-4930-a31a-0381cd077c4a","Type":"ContainerStarted","Data":"521be7fc787c1c8c959bcbea9b41a4de5e67fc662820edded8a8c75ca7e10f81"} Oct 02 11:10:12 crc kubenswrapper[4751]: I1002 11:10:12.757327 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" event={"ID":"fa36b949-67d9-4930-a31a-0381cd077c4a","Type":"ContainerStarted","Data":"a4e43ef5b49b919ccbe3219f9a80ecee160fc7f8fd7e2fec26a4b0fe9b34e74b"} Oct 02 11:10:12 crc kubenswrapper[4751]: I1002 11:10:12.757934 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:12 crc kubenswrapper[4751]: I1002 11:10:12.798541 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" podStartSLOduration=1.5870364829999999 podStartE2EDuration="9.79851697s" podCreationTimestamp="2025-10-02 11:10:03 +0000 UTC" firstStartedPulling="2025-10-02 11:10:04.214917689 +0000 UTC m=+1086.269144149" lastFinishedPulling="2025-10-02 11:10:12.426398186 +0000 UTC m=+1094.480624636" observedRunningTime="2025-10-02 11:10:12.79403105 +0000 UTC m=+1094.848257580" watchObservedRunningTime="2025-10-02 11:10:12.79851697 +0000 UTC m=+1094.852743440" Oct 02 11:10:13 crc kubenswrapper[4751]: I1002 11:10:13.768419 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-cc764bd77-ws526" Oct 02 11:10:31 crc kubenswrapper[4751]: I1002 11:10:31.507092 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:10:31 crc kubenswrapper[4751]: I1002 11:10:31.507838 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.085542 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.086821 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.089151 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jmv5x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.095310 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.096217 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.097782 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jbc2n" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.104614 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.107724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpzv\" (UniqueName: \"kubernetes.io/projected/ae064529-2f0b-4be8-93c0-9ac3cd3ab184-kube-api-access-gqpzv\") pod \"barbican-operator-controller-manager-6ff8b75857-pm99x\" (UID: \"ae064529-2f0b-4be8-93c0-9ac3cd3ab184\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.107795 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwpcm\" (UniqueName: \"kubernetes.io/projected/3bdcf078-737b-4a4e-9f17-10a55190af5b-kube-api-access-pwpcm\") pod \"cinder-operator-controller-manager-644bddb6d8-kglk4\" (UID: \"3bdcf078-737b-4a4e-9f17-10a55190af5b\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.113940 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.128338 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.129284 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.131501 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-44bbm" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.138713 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.139713 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.148676 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4vq9v" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.148834 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.149717 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.153560 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-2wbvm" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.168459 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.175235 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.185407 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.208436 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpzv\" (UniqueName: \"kubernetes.io/projected/ae064529-2f0b-4be8-93c0-9ac3cd3ab184-kube-api-access-gqpzv\") pod \"barbican-operator-controller-manager-6ff8b75857-pm99x\" (UID: \"ae064529-2f0b-4be8-93c0-9ac3cd3ab184\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.208477 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dh9j\" (UniqueName: \"kubernetes.io/projected/546a3f39-e64b-4747-806c-80b826920e8a-kube-api-access-6dh9j\") pod \"heat-operator-controller-manager-5d889d78cf-gcqh5\" (UID: \"546a3f39-e64b-4747-806c-80b826920e8a\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.208511 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbp6c\" (UniqueName: \"kubernetes.io/projected/a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a-kube-api-access-mbp6c\") pod \"glance-operator-controller-manager-84958c4d49-clwc9\" (UID: \"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.208542 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwpcm\" (UniqueName: \"kubernetes.io/projected/3bdcf078-737b-4a4e-9f17-10a55190af5b-kube-api-access-pwpcm\") pod \"cinder-operator-controller-manager-644bddb6d8-kglk4\" (UID: \"3bdcf078-737b-4a4e-9f17-10a55190af5b\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.208598 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclsk\" (UniqueName: \"kubernetes.io/projected/e6b172bd-40be-498b-beae-119a3638c8e3-kube-api-access-gclsk\") pod \"designate-operator-controller-manager-84f4f7b77b-rnd8b\" (UID: \"e6b172bd-40be-498b-beae-119a3638c8e3\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.217982 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.218982 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.224239 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-f9lxg" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.227594 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.228753 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.234208 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.234419 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jzsr8" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.251959 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.253025 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.254030 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwpcm\" (UniqueName: \"kubernetes.io/projected/3bdcf078-737b-4a4e-9f17-10a55190af5b-kube-api-access-pwpcm\") pod \"cinder-operator-controller-manager-644bddb6d8-kglk4\" (UID: \"3bdcf078-737b-4a4e-9f17-10a55190af5b\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.257968 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.259210 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-dp7z6" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.260220 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpzv\" (UniqueName: \"kubernetes.io/projected/ae064529-2f0b-4be8-93c0-9ac3cd3ab184-kube-api-access-gqpzv\") pod \"barbican-operator-controller-manager-6ff8b75857-pm99x\" (UID: \"ae064529-2f0b-4be8-93c0-9ac3cd3ab184\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.277258 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.282646 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.296485 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.297505 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.301803 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-7g6qb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309631 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclsk\" (UniqueName: \"kubernetes.io/projected/e6b172bd-40be-498b-beae-119a3638c8e3-kube-api-access-gclsk\") pod \"designate-operator-controller-manager-84f4f7b77b-rnd8b\" (UID: \"e6b172bd-40be-498b-beae-119a3638c8e3\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309672 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wlzk\" (UniqueName: \"kubernetes.io/projected/c6649ed7-0b11-4c16-8f0f-65cd3baa462e-kube-api-access-5wlzk\") pod \"ironic-operator-controller-manager-5cd4858477-4rhhf\" (UID: \"c6649ed7-0b11-4c16-8f0f-65cd3baa462e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309702 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dh9j\" (UniqueName: \"kubernetes.io/projected/546a3f39-e64b-4747-806c-80b826920e8a-kube-api-access-6dh9j\") pod \"heat-operator-controller-manager-5d889d78cf-gcqh5\" (UID: \"546a3f39-e64b-4747-806c-80b826920e8a\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309740 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9924s\" (UniqueName: \"kubernetes.io/projected/f512a98c-7304-4899-8382-65cab60c17a3-kube-api-access-9924s\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309755 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvjx2\" (UniqueName: \"kubernetes.io/projected/aec887c3-910f-4278-8834-afcc4c9b6aa3-kube-api-access-dvjx2\") pod \"keystone-operator-controller-manager-5bd55b4bff-6lcbb\" (UID: \"aec887c3-910f-4278-8834-afcc4c9b6aa3\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309784 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbp6c\" (UniqueName: \"kubernetes.io/projected/a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a-kube-api-access-mbp6c\") pod \"glance-operator-controller-manager-84958c4d49-clwc9\" (UID: \"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309813 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4kt9\" (UniqueName: \"kubernetes.io/projected/bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0-kube-api-access-q4kt9\") pod \"horizon-operator-controller-manager-9f4696d94-8jkm7\" (UID: \"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.309858 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.325542 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.329249 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.330149 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.333863 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-9hz46" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.336542 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-vq88r"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.337568 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.340943 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-47n2v" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.347782 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dh9j\" (UniqueName: \"kubernetes.io/projected/546a3f39-e64b-4747-806c-80b826920e8a-kube-api-access-6dh9j\") pod \"heat-operator-controller-manager-5d889d78cf-gcqh5\" (UID: \"546a3f39-e64b-4747-806c-80b826920e8a\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.363413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbp6c\" (UniqueName: \"kubernetes.io/projected/a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a-kube-api-access-mbp6c\") pod \"glance-operator-controller-manager-84958c4d49-clwc9\" (UID: \"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.363884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclsk\" (UniqueName: \"kubernetes.io/projected/e6b172bd-40be-498b-beae-119a3638c8e3-kube-api-access-gclsk\") pod \"designate-operator-controller-manager-84f4f7b77b-rnd8b\" (UID: \"e6b172bd-40be-498b-beae-119a3638c8e3\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.368004 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.375034 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-vq88r"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.384118 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.387054 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.396844 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.397841 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.399566 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7n8tt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.403735 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.408205 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qv7zf" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.408832 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.412237 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9924s\" (UniqueName: \"kubernetes.io/projected/f512a98c-7304-4899-8382-65cab60c17a3-kube-api-access-9924s\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.412283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvjx2\" (UniqueName: \"kubernetes.io/projected/aec887c3-910f-4278-8834-afcc4c9b6aa3-kube-api-access-dvjx2\") pod \"keystone-operator-controller-manager-5bd55b4bff-6lcbb\" (UID: \"aec887c3-910f-4278-8834-afcc4c9b6aa3\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.412352 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4kt9\" (UniqueName: \"kubernetes.io/projected/bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0-kube-api-access-q4kt9\") pod \"horizon-operator-controller-manager-9f4696d94-8jkm7\" (UID: \"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.412395 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.412422 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wlzk\" (UniqueName: \"kubernetes.io/projected/c6649ed7-0b11-4c16-8f0f-65cd3baa462e-kube-api-access-5wlzk\") pod \"ironic-operator-controller-manager-5cd4858477-4rhhf\" (UID: \"c6649ed7-0b11-4c16-8f0f-65cd3baa462e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:10:52 crc kubenswrapper[4751]: E1002 11:10:52.413068 4751 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 11:10:52 crc kubenswrapper[4751]: E1002 11:10:52.413460 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert podName:f512a98c-7304-4899-8382-65cab60c17a3 nodeName:}" failed. No retries permitted until 2025-10-02 11:10:52.913432695 +0000 UTC m=+1134.967659165 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert") pod "infra-operator-controller-manager-9d6c5db85-gqqxn" (UID: "f512a98c-7304-4899-8382-65cab60c17a3") : secret "infra-operator-webhook-server-cert" not found Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.421620 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.424938 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.430136 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9924s\" (UniqueName: \"kubernetes.io/projected/f512a98c-7304-4899-8382-65cab60c17a3-kube-api-access-9924s\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.439799 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wlzk\" (UniqueName: \"kubernetes.io/projected/c6649ed7-0b11-4c16-8f0f-65cd3baa462e-kube-api-access-5wlzk\") pod \"ironic-operator-controller-manager-5cd4858477-4rhhf\" (UID: \"c6649ed7-0b11-4c16-8f0f-65cd3baa462e\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.447643 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvjx2\" (UniqueName: \"kubernetes.io/projected/aec887c3-910f-4278-8834-afcc4c9b6aa3-kube-api-access-dvjx2\") pod \"keystone-operator-controller-manager-5bd55b4bff-6lcbb\" (UID: \"aec887c3-910f-4278-8834-afcc4c9b6aa3\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.454629 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.457254 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.463298 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.466265 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4kt9\" (UniqueName: \"kubernetes.io/projected/bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0-kube-api-access-q4kt9\") pod \"horizon-operator-controller-manager-9f4696d94-8jkm7\" (UID: \"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.477836 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.479541 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.482028 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.486000 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5n57w" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.491942 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.493489 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.495425 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.496286 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hzwp8" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.502694 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.504323 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.506840 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-bwjtg" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.513609 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x224\" (UniqueName: \"kubernetes.io/projected/f8b1c8b4-b491-44aa-885e-c6215289cb0d-kube-api-access-6x224\") pod \"nova-operator-controller-manager-64cd67b5cb-hcnpc\" (UID: \"f8b1c8b4-b491-44aa-885e-c6215289cb0d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.513682 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpdsm\" (UniqueName: \"kubernetes.io/projected/2ed3e86e-2313-4efa-987a-d8209ce59d73-kube-api-access-rpdsm\") pod \"mariadb-operator-controller-manager-88c7-vq88r\" (UID: \"2ed3e86e-2313-4efa-987a-d8209ce59d73\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.514656 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c54pw\" (UniqueName: \"kubernetes.io/projected/36f42236-589e-45ef-a69c-753d38fd54d6-kube-api-access-c54pw\") pod \"manila-operator-controller-manager-6d68dbc695-jk8xr\" (UID: \"36f42236-589e-45ef-a69c-753d38fd54d6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.514751 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2l9c\" (UniqueName: \"kubernetes.io/projected/4905929c-80af-4de9-b7c7-4ba9f3b938ad-kube-api-access-h2l9c\") pod \"neutron-operator-controller-manager-849d5b9b84-kqkzd\" (UID: \"4905929c-80af-4de9-b7c7-4ba9f3b938ad\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.525668 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.530774 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.542602 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.542812 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.545482 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.557840 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-666mh" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.568246 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.575670 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.580757 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.583073 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-llwb9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.584960 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.608524 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.610392 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.611698 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.613641 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-b5hfq" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620644 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwx98\" (UniqueName: \"kubernetes.io/projected/58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa-kube-api-access-lwx98\") pod \"octavia-operator-controller-manager-7b787867f4-vrndj\" (UID: \"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620689 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x224\" (UniqueName: \"kubernetes.io/projected/f8b1c8b4-b491-44aa-885e-c6215289cb0d-kube-api-access-6x224\") pod \"nova-operator-controller-manager-64cd67b5cb-hcnpc\" (UID: \"f8b1c8b4-b491-44aa-885e-c6215289cb0d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620733 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpdsm\" (UniqueName: \"kubernetes.io/projected/2ed3e86e-2313-4efa-987a-d8209ce59d73-kube-api-access-rpdsm\") pod \"mariadb-operator-controller-manager-88c7-vq88r\" (UID: \"2ed3e86e-2313-4efa-987a-d8209ce59d73\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620767 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6drfd\" (UniqueName: \"kubernetes.io/projected/64d6192e-f762-40ea-8d3f-9a2f8f27144b-kube-api-access-6drfd\") pod \"ovn-operator-controller-manager-9976ff44c-flblb\" (UID: \"64d6192e-f762-40ea-8d3f-9a2f8f27144b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620794 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c54pw\" (UniqueName: \"kubernetes.io/projected/36f42236-589e-45ef-a69c-753d38fd54d6-kube-api-access-c54pw\") pod \"manila-operator-controller-manager-6d68dbc695-jk8xr\" (UID: \"36f42236-589e-45ef-a69c-753d38fd54d6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620818 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620843 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg87t\" (UniqueName: \"kubernetes.io/projected/787aba9c-ec9d-4d98-bae0-77048675a98b-kube-api-access-mg87t\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.620871 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2l9c\" (UniqueName: \"kubernetes.io/projected/4905929c-80af-4de9-b7c7-4ba9f3b938ad-kube-api-access-h2l9c\") pod \"neutron-operator-controller-manager-849d5b9b84-kqkzd\" (UID: \"4905929c-80af-4de9-b7c7-4ba9f3b938ad\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.633456 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.647818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x224\" (UniqueName: \"kubernetes.io/projected/f8b1c8b4-b491-44aa-885e-c6215289cb0d-kube-api-access-6x224\") pod \"nova-operator-controller-manager-64cd67b5cb-hcnpc\" (UID: \"f8b1c8b4-b491-44aa-885e-c6215289cb0d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.649485 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-q2bvx"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.657805 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c54pw\" (UniqueName: \"kubernetes.io/projected/36f42236-589e-45ef-a69c-753d38fd54d6-kube-api-access-c54pw\") pod \"manila-operator-controller-manager-6d68dbc695-jk8xr\" (UID: \"36f42236-589e-45ef-a69c-753d38fd54d6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.658298 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-q2bvx"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.658397 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.662044 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-k4f7f" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.664724 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2l9c\" (UniqueName: \"kubernetes.io/projected/4905929c-80af-4de9-b7c7-4ba9f3b938ad-kube-api-access-h2l9c\") pod \"neutron-operator-controller-manager-849d5b9b84-kqkzd\" (UID: \"4905929c-80af-4de9-b7c7-4ba9f3b938ad\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.669913 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.674074 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpdsm\" (UniqueName: \"kubernetes.io/projected/2ed3e86e-2313-4efa-987a-d8209ce59d73-kube-api-access-rpdsm\") pod \"mariadb-operator-controller-manager-88c7-vq88r\" (UID: \"2ed3e86e-2313-4efa-987a-d8209ce59d73\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.711962 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.716350 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.717636 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722279 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722338 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg87t\" (UniqueName: \"kubernetes.io/projected/787aba9c-ec9d-4d98-bae0-77048675a98b-kube-api-access-mg87t\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722387 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtcfb\" (UniqueName: \"kubernetes.io/projected/42842b0c-bdf5-4b32-8c6d-f635c333c1b7-kube-api-access-gtcfb\") pod \"telemetry-operator-controller-manager-b8d54b5d7-pzqdp\" (UID: \"42842b0c-bdf5-4b32-8c6d-f635c333c1b7\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722428 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwx98\" (UniqueName: \"kubernetes.io/projected/58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa-kube-api-access-lwx98\") pod \"octavia-operator-controller-manager-7b787867f4-vrndj\" (UID: \"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722446 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbd7\" (UniqueName: \"kubernetes.io/projected/91ad0328-73e0-4cf6-89d0-354d79055a1b-kube-api-access-hkbd7\") pod \"swift-operator-controller-manager-84d6b4b759-cqsln\" (UID: \"91ad0328-73e0-4cf6-89d0-354d79055a1b\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722477 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kknzb\" (UniqueName: \"kubernetes.io/projected/738b4521-cee5-4fa7-95df-3ac0ea68b26f-kube-api-access-kknzb\") pod \"placement-operator-controller-manager-589c58c6c-j85jt\" (UID: \"738b4521-cee5-4fa7-95df-3ac0ea68b26f\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.722514 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6drfd\" (UniqueName: \"kubernetes.io/projected/64d6192e-f762-40ea-8d3f-9a2f8f27144b-kube-api-access-6drfd\") pod \"ovn-operator-controller-manager-9976ff44c-flblb\" (UID: \"64d6192e-f762-40ea-8d3f-9a2f8f27144b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.723715 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-96km5" Oct 02 11:10:52 crc kubenswrapper[4751]: E1002 11:10:52.724087 4751 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 11:10:52 crc kubenswrapper[4751]: E1002 11:10:52.724132 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert podName:787aba9c-ec9d-4d98-bae0-77048675a98b nodeName:}" failed. No retries permitted until 2025-10-02 11:10:53.22411609 +0000 UTC m=+1135.278342540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" (UID: "787aba9c-ec9d-4d98-bae0-77048675a98b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.735922 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.745196 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6drfd\" (UniqueName: \"kubernetes.io/projected/64d6192e-f762-40ea-8d3f-9a2f8f27144b-kube-api-access-6drfd\") pod \"ovn-operator-controller-manager-9976ff44c-flblb\" (UID: \"64d6192e-f762-40ea-8d3f-9a2f8f27144b\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.750537 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.756874 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg87t\" (UniqueName: \"kubernetes.io/projected/787aba9c-ec9d-4d98-bae0-77048675a98b-kube-api-access-mg87t\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.759503 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwx98\" (UniqueName: \"kubernetes.io/projected/58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa-kube-api-access-lwx98\") pod \"octavia-operator-controller-manager-7b787867f4-vrndj\" (UID: \"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.789510 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.807412 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.811215 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.814151 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.817734 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9w8k4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.817956 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.823249 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbd7\" (UniqueName: \"kubernetes.io/projected/91ad0328-73e0-4cf6-89d0-354d79055a1b-kube-api-access-hkbd7\") pod \"swift-operator-controller-manager-84d6b4b759-cqsln\" (UID: \"91ad0328-73e0-4cf6-89d0-354d79055a1b\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.823335 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kknzb\" (UniqueName: \"kubernetes.io/projected/738b4521-cee5-4fa7-95df-3ac0ea68b26f-kube-api-access-kknzb\") pod \"placement-operator-controller-manager-589c58c6c-j85jt\" (UID: \"738b4521-cee5-4fa7-95df-3ac0ea68b26f\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.823411 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7xjs\" (UniqueName: \"kubernetes.io/projected/a1b22cc9-2f48-475c-9325-fca7fbdfbc25-kube-api-access-t7xjs\") pod \"watcher-operator-controller-manager-6b9957f54f-st8lx\" (UID: \"a1b22cc9-2f48-475c-9325-fca7fbdfbc25\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.823440 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n4qf\" (UniqueName: \"kubernetes.io/projected/77ffa158-937f-44d2-b65a-9f448a064288-kube-api-access-7n4qf\") pod \"test-operator-controller-manager-85777745bb-q2bvx\" (UID: \"77ffa158-937f-44d2-b65a-9f448a064288\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.823466 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtcfb\" (UniqueName: \"kubernetes.io/projected/42842b0c-bdf5-4b32-8c6d-f635c333c1b7-kube-api-access-gtcfb\") pod \"telemetry-operator-controller-manager-b8d54b5d7-pzqdp\" (UID: \"42842b0c-bdf5-4b32-8c6d-f635c333c1b7\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.831230 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.831245 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.847631 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kknzb\" (UniqueName: \"kubernetes.io/projected/738b4521-cee5-4fa7-95df-3ac0ea68b26f-kube-api-access-kknzb\") pod \"placement-operator-controller-manager-589c58c6c-j85jt\" (UID: \"738b4521-cee5-4fa7-95df-3ac0ea68b26f\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.850867 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbd7\" (UniqueName: \"kubernetes.io/projected/91ad0328-73e0-4cf6-89d0-354d79055a1b-kube-api-access-hkbd7\") pod \"swift-operator-controller-manager-84d6b4b759-cqsln\" (UID: \"91ad0328-73e0-4cf6-89d0-354d79055a1b\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.852285 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtcfb\" (UniqueName: \"kubernetes.io/projected/42842b0c-bdf5-4b32-8c6d-f635c333c1b7-kube-api-access-gtcfb\") pod \"telemetry-operator-controller-manager-b8d54b5d7-pzqdp\" (UID: \"42842b0c-bdf5-4b32-8c6d-f635c333c1b7\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.937247 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.937841 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.940531 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.940600 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.940666 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbrmt\" (UniqueName: \"kubernetes.io/projected/5be43ec7-c21c-4a6c-a3ef-a816569fda32-kube-api-access-dbrmt\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.940698 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7xjs\" (UniqueName: \"kubernetes.io/projected/a1b22cc9-2f48-475c-9325-fca7fbdfbc25-kube-api-access-t7xjs\") pod \"watcher-operator-controller-manager-6b9957f54f-st8lx\" (UID: \"a1b22cc9-2f48-475c-9325-fca7fbdfbc25\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.940722 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n4qf\" (UniqueName: \"kubernetes.io/projected/77ffa158-937f-44d2-b65a-9f448a064288-kube-api-access-7n4qf\") pod \"test-operator-controller-manager-85777745bb-q2bvx\" (UID: \"77ffa158-937f-44d2-b65a-9f448a064288\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.950368 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4"] Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.951992 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.952075 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f512a98c-7304-4899-8382-65cab60c17a3-cert\") pod \"infra-operator-controller-manager-9d6c5db85-gqqxn\" (UID: \"f512a98c-7304-4899-8382-65cab60c17a3\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.952212 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.959315 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qmgn9" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.960563 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:10:52 crc kubenswrapper[4751]: I1002 11:10:52.968452 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n4qf\" (UniqueName: \"kubernetes.io/projected/77ffa158-937f-44d2-b65a-9f448a064288-kube-api-access-7n4qf\") pod \"test-operator-controller-manager-85777745bb-q2bvx\" (UID: \"77ffa158-937f-44d2-b65a-9f448a064288\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.006664 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.009908 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7xjs\" (UniqueName: \"kubernetes.io/projected/a1b22cc9-2f48-475c-9325-fca7fbdfbc25-kube-api-access-t7xjs\") pod \"watcher-operator-controller-manager-6b9957f54f-st8lx\" (UID: \"a1b22cc9-2f48-475c-9325-fca7fbdfbc25\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.021271 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.063485 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.064104 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.064289 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbrmt\" (UniqueName: \"kubernetes.io/projected/5be43ec7-c21c-4a6c-a3ef-a816569fda32-kube-api-access-dbrmt\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.064330 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh94t\" (UniqueName: \"kubernetes.io/projected/5a9de164-5d56-4ff5-ac4a-2cf16846bb4c-kube-api-access-bh94t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4\" (UID: \"5a9de164-5d56-4ff5-ac4a-2cf16846bb4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.064478 4751 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.064538 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert podName:5be43ec7-c21c-4a6c-a3ef-a816569fda32 nodeName:}" failed. No retries permitted until 2025-10-02 11:10:53.564519119 +0000 UTC m=+1135.618745569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert") pod "openstack-operator-controller-manager-5f7d749dc7-j6vzk" (UID: "5be43ec7-c21c-4a6c-a3ef-a816569fda32") : secret "webhook-server-cert" not found Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.068839 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.083884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbrmt\" (UniqueName: \"kubernetes.io/projected/5be43ec7-c21c-4a6c-a3ef-a816569fda32-kube-api-access-dbrmt\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.117994 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.136854 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.150077 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.170728 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh94t\" (UniqueName: \"kubernetes.io/projected/5a9de164-5d56-4ff5-ac4a-2cf16846bb4c-kube-api-access-bh94t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4\" (UID: \"5a9de164-5d56-4ff5-ac4a-2cf16846bb4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.187022 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh94t\" (UniqueName: \"kubernetes.io/projected/5a9de164-5d56-4ff5-ac4a-2cf16846bb4c-kube-api-access-bh94t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4\" (UID: \"5a9de164-5d56-4ff5-ac4a-2cf16846bb4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.208595 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.232014 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bdcf078_737b_4a4e_9f17_10a55190af5b.slice/crio-48dbdbf6e660e094f07f033ea59dfa7e285483dace926cab2fdc2e85b8aceacf WatchSource:0}: Error finding container 48dbdbf6e660e094f07f033ea59dfa7e285483dace926cab2fdc2e85b8aceacf: Status 404 returned error can't find the container with id 48dbdbf6e660e094f07f033ea59dfa7e285483dace926cab2fdc2e85b8aceacf Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.245469 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.272435 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.272690 4751 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.272749 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert podName:787aba9c-ec9d-4d98-bae0-77048675a98b nodeName:}" failed. No retries permitted until 2025-10-02 11:10:54.272731165 +0000 UTC m=+1136.326957615 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" (UID: "787aba9c-ec9d-4d98-bae0-77048675a98b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.304018 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.367055 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.583757 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.588497 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5be43ec7-c21c-4a6c-a3ef-a816569fda32-cert\") pod \"openstack-operator-controller-manager-5f7d749dc7-j6vzk\" (UID: \"5be43ec7-c21c-4a6c-a3ef-a816569fda32\") " pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.592905 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd"] Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.597092 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6649ed7_0b11_4c16_8f0f_65cd3baa462e.slice/crio-f5437e0455dc6f2c9d0be1756a89a00a7e4dee14cabe83ad2fa5a68b43726c91 WatchSource:0}: Error finding container f5437e0455dc6f2c9d0be1756a89a00a7e4dee14cabe83ad2fa5a68b43726c91: Status 404 returned error can't find the container with id f5437e0455dc6f2c9d0be1756a89a00a7e4dee14cabe83ad2fa5a68b43726c91 Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.599229 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb"] Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.603768 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaec887c3_910f_4278_8834_afcc4c9b6aa3.slice/crio-8a9eaa3e99c602b173299186f59e1b262da65bb999052047091b2050138d4887 WatchSource:0}: Error finding container 8a9eaa3e99c602b173299186f59e1b262da65bb999052047091b2050138d4887: Status 404 returned error can't find the container with id 8a9eaa3e99c602b173299186f59e1b262da65bb999052047091b2050138d4887 Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.607957 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.730068 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.737646 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-vq88r"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.741949 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.750441 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.761220 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc"] Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.776598 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8b1c8b4_b491_44aa_885e_c6215289cb0d.slice/crio-82b6233a723d316c6b5d2c23dab787ce72f9a42751feeb633f6775e602c40513 WatchSource:0}: Error finding container 82b6233a723d316c6b5d2c23dab787ce72f9a42751feeb633f6775e602c40513: Status 404 returned error can't find the container with id 82b6233a723d316c6b5d2c23dab787ce72f9a42751feeb633f6775e602c40513 Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.870665 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.874659 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln"] Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.882091 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkbd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-cqsln_openstack-operators(91ad0328-73e0-4cf6-89d0-354d79055a1b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.887124 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtcfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-pzqdp_openstack-operators(42842b0c-bdf5-4b32-8c6d-f635c333c1b7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.908560 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.913512 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-q2bvx"] Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.921586 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb"] Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.923725 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod738b4521_cee5_4fa7_95df_3ac0ea68b26f.slice/crio-b4a6eedd9de458e8c1dd5d8976028ac6aa0583671ac18dd9ddbd14c289dc2968 WatchSource:0}: Error finding container b4a6eedd9de458e8c1dd5d8976028ac6aa0583671ac18dd9ddbd14c289dc2968: Status 404 returned error can't find the container with id b4a6eedd9de458e8c1dd5d8976028ac6aa0583671ac18dd9ddbd14c289dc2968 Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.925937 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77ffa158_937f_44d2_b65a_9f448a064288.slice/crio-75ae4ffe98338fe71d4b984a562b59cd18e696657130c002f7b8210cdfae0337 WatchSource:0}: Error finding container 75ae4ffe98338fe71d4b984a562b59cd18e696657130c002f7b8210cdfae0337: Status 404 returned error can't find the container with id 75ae4ffe98338fe71d4b984a562b59cd18e696657130c002f7b8210cdfae0337 Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.935660 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx"] Oct 02 11:10:53 crc kubenswrapper[4751]: W1002 11:10:53.946395 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64d6192e_f762_40ea_8d3f_9a2f8f27144b.slice/crio-8299fa523ec14139349d5660a235efec9238c7fb659aea15f7832ade375ebdde WatchSource:0}: Error finding container 8299fa523ec14139349d5660a235efec9238c7fb659aea15f7832ade375ebdde: Status 404 returned error can't find the container with id 8299fa523ec14139349d5660a235efec9238c7fb659aea15f7832ade375ebdde Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.946585 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kknzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-j85jt_openstack-operators(738b4521-cee5-4fa7-95df-3ac0ea68b26f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.946638 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n4qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-q2bvx_openstack-operators(77ffa158-937f-44d2-b65a-9f448a064288): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.953812 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6drfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-flblb_openstack-operators(64d6192e-f762-40ea-8d3f-9a2f8f27144b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: E1002 11:10:53.952068 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t7xjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6b9957f54f-st8lx_openstack-operators(a1b22cc9-2f48-475c-9325-fca7fbdfbc25): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:53 crc kubenswrapper[4751]: I1002 11:10:53.982144 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk"] Oct 02 11:10:54 crc kubenswrapper[4751]: W1002 11:10:54.012277 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5be43ec7_c21c_4a6c_a3ef_a816569fda32.slice/crio-963fb88d0e646e01878c8b88244f8b7b1904ed2f54d85232828b05ca8765b8e7 WatchSource:0}: Error finding container 963fb88d0e646e01878c8b88244f8b7b1904ed2f54d85232828b05ca8765b8e7: Status 404 returned error can't find the container with id 963fb88d0e646e01878c8b88244f8b7b1904ed2f54d85232828b05ca8765b8e7 Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.027236 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4"] Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.050821 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn"] Oct 02 11:10:54 crc kubenswrapper[4751]: W1002 11:10:54.069275 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a9de164_5d56_4ff5_ac4a_2cf16846bb4c.slice/crio-6ab4d220fcc0e6c9916413d920d2924e5a938391afbd3c39b702c339f34bc467 WatchSource:0}: Error finding container 6ab4d220fcc0e6c9916413d920d2924e5a938391afbd3c39b702c339f34bc467: Status 404 returned error can't find the container with id 6ab4d220fcc0e6c9916413d920d2924e5a938391afbd3c39b702c339f34bc467 Oct 02 11:10:54 crc kubenswrapper[4751]: W1002 11:10:54.071056 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf512a98c_7304_4899_8382_65cab60c17a3.slice/crio-22432726ed4c59c284b8c73036011b30fb673836b3c48f2c150cf7895ae11e63 WatchSource:0}: Error finding container 22432726ed4c59c284b8c73036011b30fb673836b3c48f2c150cf7895ae11e63: Status 404 returned error can't find the container with id 22432726ed4c59c284b8c73036011b30fb673836b3c48f2c150cf7895ae11e63 Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.076050 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" event={"ID":"ae064529-2f0b-4be8-93c0-9ac3cd3ab184","Type":"ContainerStarted","Data":"c3aaa2bcbbbe2a7d0a8e248ba2a5ca4eb43e256610b2737212ae17295aad28c2"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.077404 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" event={"ID":"a1b22cc9-2f48-475c-9325-fca7fbdfbc25","Type":"ContainerStarted","Data":"3016adcd6f9acb5f8617b57c992667a51d15599a089d1ecbebcacbbe7aa2d2fd"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.079005 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" event={"ID":"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa","Type":"ContainerStarted","Data":"8895d5e6d31998110e49f62798f6ff96a469ad480d1839258ce6e50cfd9a43c5"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.082511 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" event={"ID":"42842b0c-bdf5-4b32-8c6d-f635c333c1b7","Type":"ContainerStarted","Data":"cb9c7ca565ee82b03cab14e6004933e78a72c11b3bcecea55ee568e8a131b6e1"} Oct 02 11:10:54 crc kubenswrapper[4751]: E1002 11:10:54.082622 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9924s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-9d6c5db85-gqqxn_openstack-operators(f512a98c-7304-4899-8382-65cab60c17a3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.086488 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" event={"ID":"3bdcf078-737b-4a4e-9f17-10a55190af5b","Type":"ContainerStarted","Data":"48dbdbf6e660e094f07f033ea59dfa7e285483dace926cab2fdc2e85b8aceacf"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.088565 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" event={"ID":"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a","Type":"ContainerStarted","Data":"4da0104b39a3a78be4d2648471b224412f554f55e92b5c85c65927beb416ca48"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.089569 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" event={"ID":"5be43ec7-c21c-4a6c-a3ef-a816569fda32","Type":"ContainerStarted","Data":"963fb88d0e646e01878c8b88244f8b7b1904ed2f54d85232828b05ca8765b8e7"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.090593 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" event={"ID":"2ed3e86e-2313-4efa-987a-d8209ce59d73","Type":"ContainerStarted","Data":"292e7c5a65f2a5be5007e013f3d723a44e6a0adfd7565f3c69ab9ecc25079267"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.093937 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" event={"ID":"e6b172bd-40be-498b-beae-119a3638c8e3","Type":"ContainerStarted","Data":"ab0b4eaa8b9b1e46aa343f0df8871246094980dd69b6e4f64a8dde606b8548a1"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.095468 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" event={"ID":"738b4521-cee5-4fa7-95df-3ac0ea68b26f","Type":"ContainerStarted","Data":"b4a6eedd9de458e8c1dd5d8976028ac6aa0583671ac18dd9ddbd14c289dc2968"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.097581 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" event={"ID":"f8b1c8b4-b491-44aa-885e-c6215289cb0d","Type":"ContainerStarted","Data":"82b6233a723d316c6b5d2c23dab787ce72f9a42751feeb633f6775e602c40513"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.099373 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" event={"ID":"36f42236-589e-45ef-a69c-753d38fd54d6","Type":"ContainerStarted","Data":"2d29f841e01c57fcfff0709ff06d026d3ee5cb0920abc056bd74313cff134e8f"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.100646 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" event={"ID":"4905929c-80af-4de9-b7c7-4ba9f3b938ad","Type":"ContainerStarted","Data":"207e1b82e491eb742eabf52e9ff690ff77bb155e22805629ef078a6621c6b09d"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.102090 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" event={"ID":"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0","Type":"ContainerStarted","Data":"90a8c9ed5610bcf81fc9130b826e6efcef0a097d01c1493e8f92f90d7a7d281a"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.104482 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" event={"ID":"546a3f39-e64b-4747-806c-80b826920e8a","Type":"ContainerStarted","Data":"8723880fcf3dee7b97d9c4a71399f7bbad42eb478354a2c9c83489400517c9f1"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.105781 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" event={"ID":"64d6192e-f762-40ea-8d3f-9a2f8f27144b","Type":"ContainerStarted","Data":"8299fa523ec14139349d5660a235efec9238c7fb659aea15f7832ade375ebdde"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.115921 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" event={"ID":"77ffa158-937f-44d2-b65a-9f448a064288","Type":"ContainerStarted","Data":"75ae4ffe98338fe71d4b984a562b59cd18e696657130c002f7b8210cdfae0337"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.117398 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" event={"ID":"91ad0328-73e0-4cf6-89d0-354d79055a1b","Type":"ContainerStarted","Data":"db36620ffc22a43ab8b46ada2991570de62a61e8249edd381f52bcc651bd627c"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.119001 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" event={"ID":"c6649ed7-0b11-4c16-8f0f-65cd3baa462e","Type":"ContainerStarted","Data":"f5437e0455dc6f2c9d0be1756a89a00a7e4dee14cabe83ad2fa5a68b43726c91"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.120139 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" event={"ID":"aec887c3-910f-4278-8834-afcc4c9b6aa3","Type":"ContainerStarted","Data":"8a9eaa3e99c602b173299186f59e1b262da65bb999052047091b2050138d4887"} Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.297531 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.304473 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/787aba9c-ec9d-4d98-bae0-77048675a98b-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-h7qc4\" (UID: \"787aba9c-ec9d-4d98-bae0-77048675a98b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.344608 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:10:54 crc kubenswrapper[4751]: I1002 11:10:54.572758 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4"] Oct 02 11:10:54 crc kubenswrapper[4751]: E1002 11:10:54.851112 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podUID="42842b0c-bdf5-4b32-8c6d-f635c333c1b7" Oct 02 11:10:54 crc kubenswrapper[4751]: E1002 11:10:54.852851 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" podUID="91ad0328-73e0-4cf6-89d0-354d79055a1b" Oct 02 11:10:54 crc kubenswrapper[4751]: E1002 11:10:54.927790 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podUID="77ffa158-937f-44d2-b65a-9f448a064288" Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.000790 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" podUID="a1b22cc9-2f48-475c-9325-fca7fbdfbc25" Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.004235 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podUID="64d6192e-f762-40ea-8d3f-9a2f8f27144b" Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.006633 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" podUID="f512a98c-7304-4899-8382-65cab60c17a3" Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.097088 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podUID="738b4521-cee5-4fa7-95df-3ac0ea68b26f" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.195455 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" event={"ID":"f512a98c-7304-4899-8382-65cab60c17a3","Type":"ContainerStarted","Data":"78af15583272fac0fd340b3aacb40554b5242d43474a3f10ac6f599824e051ad"} Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.195502 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" event={"ID":"f512a98c-7304-4899-8382-65cab60c17a3","Type":"ContainerStarted","Data":"22432726ed4c59c284b8c73036011b30fb673836b3c48f2c150cf7895ae11e63"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.205964 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" podUID="f512a98c-7304-4899-8382-65cab60c17a3" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.270329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" event={"ID":"787aba9c-ec9d-4d98-bae0-77048675a98b","Type":"ContainerStarted","Data":"858a1dda3d80e406e96dd68e46fd658ac0c4623040cbb00815833a536e58d5cd"} Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.281850 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" event={"ID":"42842b0c-bdf5-4b32-8c6d-f635c333c1b7","Type":"ContainerStarted","Data":"f7ce5d195c5eace25a480284827c3f691632cbc42c4d94a884bba0df7e0aa44c"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.284570 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podUID="42842b0c-bdf5-4b32-8c6d-f635c333c1b7" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.306559 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" event={"ID":"738b4521-cee5-4fa7-95df-3ac0ea68b26f","Type":"ContainerStarted","Data":"b741c09cce43e6af3518a9014a7c1ee15fe2419c8842d872f1354b8b38055eef"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.308013 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podUID="738b4521-cee5-4fa7-95df-3ac0ea68b26f" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.324491 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" event={"ID":"5be43ec7-c21c-4a6c-a3ef-a816569fda32","Type":"ContainerStarted","Data":"8448a1dae489639dc46bdc016f896a59680b0d0dcee673dd92b4e4c3b87fc282"} Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.333695 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" event={"ID":"64d6192e-f762-40ea-8d3f-9a2f8f27144b","Type":"ContainerStarted","Data":"a1e3d0501dc58a9339802a5c0fecae077a18d8afbf250179baf2aed80821a47e"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.337087 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podUID="64d6192e-f762-40ea-8d3f-9a2f8f27144b" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.338949 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" event={"ID":"a1b22cc9-2f48-475c-9325-fca7fbdfbc25","Type":"ContainerStarted","Data":"625352c397586ca09fbad34357c7c12a34d4e70767bb1ab954c598cb60163f70"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.363068 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" podUID="a1b22cc9-2f48-475c-9325-fca7fbdfbc25" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.365859 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" event={"ID":"77ffa158-937f-44d2-b65a-9f448a064288","Type":"ContainerStarted","Data":"fd676a5fc10eddaa0f9c52f82408a7fd0c621ae7f4289797626f7f99208d26d5"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.367990 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podUID="77ffa158-937f-44d2-b65a-9f448a064288" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.384493 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" event={"ID":"91ad0328-73e0-4cf6-89d0-354d79055a1b","Type":"ContainerStarted","Data":"94964041bc01b39a199f5d5781be1742d52127c5d450eb8678b7ab7fca87c6e4"} Oct 02 11:10:55 crc kubenswrapper[4751]: E1002 11:10:55.393404 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" podUID="91ad0328-73e0-4cf6-89d0-354d79055a1b" Oct 02 11:10:55 crc kubenswrapper[4751]: I1002 11:10:55.404426 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" event={"ID":"5a9de164-5d56-4ff5-ac4a-2cf16846bb4c","Type":"ContainerStarted","Data":"6ab4d220fcc0e6c9916413d920d2924e5a938391afbd3c39b702c339f34bc467"} Oct 02 11:10:56 crc kubenswrapper[4751]: I1002 11:10:56.418451 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" event={"ID":"5be43ec7-c21c-4a6c-a3ef-a816569fda32","Type":"ContainerStarted","Data":"9bbd9aefd66ff0902c5ff5cdc2999ecdb539e080bc8290d2c1ca1a3754f19737"} Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.420408 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podUID="738b4521-cee5-4fa7-95df-3ac0ea68b26f" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.421651 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podUID="42842b0c-bdf5-4b32-8c6d-f635c333c1b7" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.421804 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podUID="77ffa158-937f-44d2-b65a-9f448a064288" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.421912 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podUID="64d6192e-f762-40ea-8d3f-9a2f8f27144b" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.422013 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" podUID="f512a98c-7304-4899-8382-65cab60c17a3" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.422122 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" podUID="a1b22cc9-2f48-475c-9325-fca7fbdfbc25" Oct 02 11:10:56 crc kubenswrapper[4751]: E1002 11:10:56.422152 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" podUID="91ad0328-73e0-4cf6-89d0-354d79055a1b" Oct 02 11:10:56 crc kubenswrapper[4751]: I1002 11:10:56.531443 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" podStartSLOduration=4.531427456 podStartE2EDuration="4.531427456s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:10:56.525246231 +0000 UTC m=+1138.579472701" watchObservedRunningTime="2025-10-02 11:10:56.531427456 +0000 UTC m=+1138.585653906" Oct 02 11:10:57 crc kubenswrapper[4751]: I1002 11:10:57.427397 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:11:01 crc kubenswrapper[4751]: I1002 11:11:01.507761 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:01 crc kubenswrapper[4751]: I1002 11:11:01.508406 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:03 crc kubenswrapper[4751]: I1002 11:11:03.762102 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5f7d749dc7-j6vzk" Oct 02 11:11:07 crc kubenswrapper[4751]: I1002 11:11:07.648253 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:11:14 crc kubenswrapper[4751]: E1002 11:11:14.377721 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a" Oct 02 11:11:14 crc kubenswrapper[4751]: E1002 11:11:14.378776 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dvjx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5bd55b4bff-6lcbb_openstack-operators(aec887c3-910f-4278-8834-afcc4c9b6aa3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:16 crc kubenswrapper[4751]: E1002 11:11:16.566148 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9" Oct 02 11:11:16 crc kubenswrapper[4751]: E1002 11:11:16.566594 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwx98,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7b787867f4-vrndj_openstack-operators(58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:18 crc kubenswrapper[4751]: E1002 11:11:18.780043 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8" Oct 02 11:11:18 crc kubenswrapper[4751]: E1002 11:11:18.780255 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h2l9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-849d5b9b84-kqkzd_openstack-operators(4905929c-80af-4de9-b7c7-4ba9f3b938ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:21 crc kubenswrapper[4751]: E1002 11:11:21.354655 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Oct 02 11:11:21 crc kubenswrapper[4751]: E1002 11:11:21.355260 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bh94t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4_openstack-operators(5a9de164-5d56-4ff5-ac4a-2cf16846bb4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:21 crc kubenswrapper[4751]: E1002 11:11:21.356542 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" podUID="5a9de164-5d56-4ff5-ac4a-2cf16846bb4c" Oct 02 11:11:21 crc kubenswrapper[4751]: E1002 11:11:21.603799 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" podUID="5a9de164-5d56-4ff5-ac4a-2cf16846bb4c" Oct 02 11:11:23 crc kubenswrapper[4751]: E1002 11:11:23.575798 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a" Oct 02 11:11:23 crc kubenswrapper[4751]: E1002 11:11:23.576998 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rpdsm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-88c7-vq88r_openstack-operators(2ed3e86e-2313-4efa-987a-d8209ce59d73): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:24 crc kubenswrapper[4751]: E1002 11:11:24.201438 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5" Oct 02 11:11:24 crc kubenswrapper[4751]: E1002 11:11:24.201762 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5wlzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5cd4858477-4rhhf_openstack-operators(c6649ed7-0b11-4c16-8f0f-65cd3baa462e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:24 crc kubenswrapper[4751]: E1002 11:11:24.519510 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72" Oct 02 11:11:24 crc kubenswrapper[4751]: E1002 11:11:24.519727 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbp6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-84958c4d49-clwc9_openstack-operators(a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:25 crc kubenswrapper[4751]: E1002 11:11:25.231840 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f" Oct 02 11:11:25 crc kubenswrapper[4751]: E1002 11:11:25.232151 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6x224,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-hcnpc_openstack-operators(f8b1c8b4-b491-44aa-885e-c6215289cb0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:26 crc kubenswrapper[4751]: E1002 11:11:26.088998 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1" Oct 02 11:11:26 crc kubenswrapper[4751]: E1002 11:11:26.089447 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pwpcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-644bddb6d8-kglk4_openstack-operators(3bdcf078-737b-4a4e-9f17-10a55190af5b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:26 crc kubenswrapper[4751]: E1002 11:11:26.650128 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884" Oct 02 11:11:26 crc kubenswrapper[4751]: E1002 11:11:26.650878 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c54pw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-jk8xr_openstack-operators(36f42236-589e-45ef-a69c-753d38fd54d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:27 crc kubenswrapper[4751]: E1002 11:11:27.082811 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8" Oct 02 11:11:27 crc kubenswrapper[4751]: E1002 11:11:27.083804 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mg87t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-h7qc4_openstack-operators(787aba9c-ec9d-4d98-bae0-77048675a98b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.507573 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.507639 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.507683 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.508438 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.508500 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182" gracePeriod=600 Oct 02 11:11:31 crc kubenswrapper[4751]: E1002 11:11:31.635357 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3" Oct 02 11:11:31 crc kubenswrapper[4751]: E1002 11:11:31.635902 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n4qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-q2bvx_openstack-operators(77ffa158-937f-44d2-b65a-9f448a064288): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:31 crc kubenswrapper[4751]: E1002 11:11:31.637373 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podUID="77ffa158-937f-44d2-b65a-9f448a064288" Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.682972 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182" exitCode=0 Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.683024 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182"} Oct 02 11:11:31 crc kubenswrapper[4751]: I1002 11:11:31.683061 4751 scope.go:117] "RemoveContainer" containerID="3feb7b30541f5940761ff8f09dd1f221290d4ee7c86857fbb28884d5b6671f0c" Oct 02 11:11:33 crc kubenswrapper[4751]: E1002 11:11:33.351061 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f" Oct 02 11:11:33 crc kubenswrapper[4751]: E1002 11:11:33.352295 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtcfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-pzqdp_openstack-operators(42842b0c-bdf5-4b32-8c6d-f635c333c1b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:33 crc kubenswrapper[4751]: E1002 11:11:33.353754 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podUID="42842b0c-bdf5-4b32-8c6d-f635c333c1b7" Oct 02 11:11:34 crc kubenswrapper[4751]: E1002 11:11:34.040142 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2" Oct 02 11:11:34 crc kubenswrapper[4751]: E1002 11:11:34.040409 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kknzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-j85jt_openstack-operators(738b4521-cee5-4fa7-95df-3ac0ea68b26f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:34 crc kubenswrapper[4751]: E1002 11:11:34.041891 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podUID="738b4521-cee5-4fa7-95df-3ac0ea68b26f" Oct 02 11:11:35 crc kubenswrapper[4751]: E1002 11:11:35.600281 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302" Oct 02 11:11:35 crc kubenswrapper[4751]: E1002 11:11:35.600707 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6drfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-flblb_openstack-operators(64d6192e-f762-40ea-8d3f-9a2f8f27144b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:11:35 crc kubenswrapper[4751]: E1002 11:11:35.602280 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podUID="64d6192e-f762-40ea-8d3f-9a2f8f27144b" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.892236 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" podUID="4905929c-80af-4de9-b7c7-4ba9f3b938ad" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.892642 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" podUID="58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.892740 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" podUID="f8b1c8b4-b491-44aa-885e-c6215289cb0d" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.893008 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" podUID="36f42236-589e-45ef-a69c-753d38fd54d6" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.893413 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" podUID="2ed3e86e-2313-4efa-987a-d8209ce59d73" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.893673 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" podUID="a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a" Oct 02 11:11:36 crc kubenswrapper[4751]: E1002 11:11:36.925761 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" podUID="c6649ed7-0b11-4c16-8f0f-65cd3baa462e" Oct 02 11:11:37 crc kubenswrapper[4751]: E1002 11:11:37.089329 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" podUID="787aba9c-ec9d-4d98-bae0-77048675a98b" Oct 02 11:11:37 crc kubenswrapper[4751]: E1002 11:11:37.089889 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" podUID="aec887c3-910f-4278-8834-afcc4c9b6aa3" Oct 02 11:11:37 crc kubenswrapper[4751]: E1002 11:11:37.090768 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" podUID="3bdcf078-737b-4a4e-9f17-10a55190af5b" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.734361 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" event={"ID":"36f42236-589e-45ef-a69c-753d38fd54d6","Type":"ContainerStarted","Data":"5d280cac91ca69fb08a8116bbc7aba8b1bd04ebebd6eb8ed86d37747117db490"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.736404 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" event={"ID":"91ad0328-73e0-4cf6-89d0-354d79055a1b","Type":"ContainerStarted","Data":"2a51eef2092f3c84abfe5a7229822fecd1ae5bf98fc4fae8d611365bf1366324"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.736571 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.738007 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" event={"ID":"aec887c3-910f-4278-8834-afcc4c9b6aa3","Type":"ContainerStarted","Data":"d9510a0a7e0f89cdd1c180accff40bce4b7b3d35e9e5bf25313ae4ac0d466e9c"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.740492 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" event={"ID":"f8b1c8b4-b491-44aa-885e-c6215289cb0d","Type":"ContainerStarted","Data":"dd1847c9bdb5c0cfd97c2cc0dc1fa16c9e3e01308c1ed8498c519c9ed9485527"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.742333 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" event={"ID":"f512a98c-7304-4899-8382-65cab60c17a3","Type":"ContainerStarted","Data":"a377443fe42307569c89d1299f6dd8ccde2e4049504e3ffce8c8daf3222aa7f4"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.742544 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.745098 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.746996 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" event={"ID":"a1b22cc9-2f48-475c-9325-fca7fbdfbc25","Type":"ContainerStarted","Data":"87182e50166c2a4b873d1499ae9f4f7a1654836f15a7da55a07563a4bd533bf7"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.747201 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.748426 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" event={"ID":"2ed3e86e-2313-4efa-987a-d8209ce59d73","Type":"ContainerStarted","Data":"31cac11434abe67683fcbaada9e603f987df063ce8b016e85324c7aff7768dce"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.749711 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" event={"ID":"4905929c-80af-4de9-b7c7-4ba9f3b938ad","Type":"ContainerStarted","Data":"91710a3633e428592ff03a2b9eb0cf03320b1bc5bcff0e03e903b4dad27bd69b"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.751372 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" event={"ID":"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa","Type":"ContainerStarted","Data":"9ee74ecd19cd00ceabc3f918f7ca779e50a5af63497e06ce8ab1a4792d76d8f8"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.758987 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" event={"ID":"787aba9c-ec9d-4d98-bae0-77048675a98b","Type":"ContainerStarted","Data":"05ab0db2f9b6625cc835f799dea63b9fe74f8125a82647cb1014755fafa406a9"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.761253 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" event={"ID":"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a","Type":"ContainerStarted","Data":"045bcbe48ac3cdd9a67db4574447aceb0ae9f91de789f4cd814395c52d35717a"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.770537 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" event={"ID":"e6b172bd-40be-498b-beae-119a3638c8e3","Type":"ContainerStarted","Data":"01643a35ee1c56d61721034cca0676f4b381358c62ff9e8d405b021a469ad3fe"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.770568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" event={"ID":"e6b172bd-40be-498b-beae-119a3638c8e3","Type":"ContainerStarted","Data":"d295e9e23d2e62cb2711991f99183739fc2107eefb4b54d5f4586ebc15ea5290"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.771092 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.773187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" event={"ID":"3bdcf078-737b-4a4e-9f17-10a55190af5b","Type":"ContainerStarted","Data":"fd3b283d46d5597a1532ca13da21c7db89d87082b78f84e6acbcab5ff3b7c23a"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.775388 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" event={"ID":"ae064529-2f0b-4be8-93c0-9ac3cd3ab184","Type":"ContainerStarted","Data":"e2ec8824f1e0bacd8cebb53015e516fa0d727068c02e1f07af7e4e8f2cfea30f"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.775429 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" event={"ID":"ae064529-2f0b-4be8-93c0-9ac3cd3ab184","Type":"ContainerStarted","Data":"dd072b6020cae71a3af9da6ed8a4b7ee6b12713a89569d737e99a8ced5e7768b"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.775467 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.777361 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" event={"ID":"c6649ed7-0b11-4c16-8f0f-65cd3baa462e","Type":"ContainerStarted","Data":"454297e6ae13930e297cc3ae394ec23a2660bd2b35fb660baf06395bcdb4b0bc"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.779355 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" event={"ID":"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0","Type":"ContainerStarted","Data":"f8d62cdd9728c9ee2bd146cf75e4571bcc007367a5d5b8fbce0c78a3cd73ec25"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.779392 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" event={"ID":"bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0","Type":"ContainerStarted","Data":"bb1af0367bf96ffc0842482ff008e562ede8c311b3331f6a7cc9d1ca487e2d11"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.779482 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.781266 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" event={"ID":"546a3f39-e64b-4747-806c-80b826920e8a","Type":"ContainerStarted","Data":"8e9ca2b4524f076510fa992230c28a6d369e0c28cfb88795eb970bbb7b36fc48"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.781303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" event={"ID":"546a3f39-e64b-4747-806c-80b826920e8a","Type":"ContainerStarted","Data":"38bb974c588fc701029f72b34e2727b10e94507d889b15ea53b4a7c2647950be"} Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.781393 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.820816 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" podStartSLOduration=4.216062995 podStartE2EDuration="45.820798919s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:54.081337041 +0000 UTC m=+1136.135563491" lastFinishedPulling="2025-10-02 11:11:35.686072965 +0000 UTC m=+1177.740299415" observedRunningTime="2025-10-02 11:11:37.820566133 +0000 UTC m=+1179.874792583" watchObservedRunningTime="2025-10-02 11:11:37.820798919 +0000 UTC m=+1179.875025369" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.844403 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" podStartSLOduration=4.052502863 podStartE2EDuration="45.844384309s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.881972372 +0000 UTC m=+1135.936198822" lastFinishedPulling="2025-10-02 11:11:35.673853808 +0000 UTC m=+1177.728080268" observedRunningTime="2025-10-02 11:11:37.837827784 +0000 UTC m=+1179.892054254" watchObservedRunningTime="2025-10-02 11:11:37.844384309 +0000 UTC m=+1179.898610759" Oct 02 11:11:37 crc kubenswrapper[4751]: I1002 11:11:37.872596 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" podStartSLOduration=4.137386852 podStartE2EDuration="45.872576923s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.950919055 +0000 UTC m=+1136.005145505" lastFinishedPulling="2025-10-02 11:11:35.686109126 +0000 UTC m=+1177.740335576" observedRunningTime="2025-10-02 11:11:37.86834023 +0000 UTC m=+1179.922566680" watchObservedRunningTime="2025-10-02 11:11:37.872576923 +0000 UTC m=+1179.926803373" Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.010839 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" podStartSLOduration=12.133826231 podStartE2EDuration="46.010825709s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.197275008 +0000 UTC m=+1135.251501458" lastFinishedPulling="2025-10-02 11:11:27.074274466 +0000 UTC m=+1169.128500936" observedRunningTime="2025-10-02 11:11:38.007980693 +0000 UTC m=+1180.062207133" watchObservedRunningTime="2025-10-02 11:11:38.010825709 +0000 UTC m=+1180.065052159" Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.027930 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" podStartSLOduration=9.078095557 podStartE2EDuration="46.027911976s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.160962648 +0000 UTC m=+1135.215189088" lastFinishedPulling="2025-10-02 11:11:30.110779017 +0000 UTC m=+1172.165005507" observedRunningTime="2025-10-02 11:11:38.022617314 +0000 UTC m=+1180.076843754" watchObservedRunningTime="2025-10-02 11:11:38.027911976 +0000 UTC m=+1180.082138426" Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.036386 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" podStartSLOduration=12.936429026999999 podStartE2EDuration="46.036371092s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.392479757 +0000 UTC m=+1135.446706207" lastFinishedPulling="2025-10-02 11:11:26.492421802 +0000 UTC m=+1168.546648272" observedRunningTime="2025-10-02 11:11:38.034593745 +0000 UTC m=+1180.088820205" watchObservedRunningTime="2025-10-02 11:11:38.036371092 +0000 UTC m=+1180.090597552" Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.082377 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" podStartSLOduration=9.335894299 podStartE2EDuration="46.082359552s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.364303143 +0000 UTC m=+1135.418529593" lastFinishedPulling="2025-10-02 11:11:30.110768356 +0000 UTC m=+1172.164994846" observedRunningTime="2025-10-02 11:11:38.069896939 +0000 UTC m=+1180.124123399" watchObservedRunningTime="2025-10-02 11:11:38.082359552 +0000 UTC m=+1180.136586002" Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.790544 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" event={"ID":"5a9de164-5d56-4ff5-ac4a-2cf16846bb4c","Type":"ContainerStarted","Data":"00a6543f7b9e66eb7109b2d6b6dd2ee74f8fd6d955b996ad1009ede9df77e8f1"} Oct 02 11:11:38 crc kubenswrapper[4751]: I1002 11:11:38.810997 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4" podStartSLOduration=2.809064515 podStartE2EDuration="46.810951619s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:54.080242352 +0000 UTC m=+1136.134468802" lastFinishedPulling="2025-10-02 11:11:38.082129456 +0000 UTC m=+1180.136355906" observedRunningTime="2025-10-02 11:11:38.809434228 +0000 UTC m=+1180.863660678" watchObservedRunningTime="2025-10-02 11:11:38.810951619 +0000 UTC m=+1180.865178089" Oct 02 11:11:42 crc kubenswrapper[4751]: I1002 11:11:42.412695 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-pm99x" Oct 02 11:11:42 crc kubenswrapper[4751]: I1002 11:11:42.461349 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rnd8b" Oct 02 11:11:42 crc kubenswrapper[4751]: I1002 11:11:42.484570 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-gcqh5" Oct 02 11:11:42 crc kubenswrapper[4751]: I1002 11:11:42.544926 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-8jkm7" Oct 02 11:11:42 crc kubenswrapper[4751]: I1002 11:11:42.941365 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-cqsln" Oct 02 11:11:43 crc kubenswrapper[4751]: I1002 11:11:43.065940 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-st8lx" Oct 02 11:11:43 crc kubenswrapper[4751]: I1002 11:11:43.216784 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-gqqxn" Oct 02 11:11:43 crc kubenswrapper[4751]: E1002 11:11:43.585147 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podUID="77ffa158-937f-44d2-b65a-9f448a064288" Oct 02 11:11:43 crc kubenswrapper[4751]: E1002 11:11:43.585452 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podUID="42842b0c-bdf5-4b32-8c6d-f635c333c1b7" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.843114 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" event={"ID":"36f42236-589e-45ef-a69c-753d38fd54d6","Type":"ContainerStarted","Data":"ed6b6ca0015fdcc219303a6d4901f4ca0a513decd19748d494380a6410fb9ab3"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.843698 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.846068 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" event={"ID":"58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa","Type":"ContainerStarted","Data":"bc92f5b62fecbc13a0610f0bfef6a11dfb20601310e89f3f377c133edfa8cafe"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.846194 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.848022 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" event={"ID":"c6649ed7-0b11-4c16-8f0f-65cd3baa462e","Type":"ContainerStarted","Data":"835f5bf467e6f12837edf00de36d93e9ca4e144e1a2a803b1b4688311372e3da"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.848064 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.849448 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" event={"ID":"aec887c3-910f-4278-8834-afcc4c9b6aa3","Type":"ContainerStarted","Data":"6a1ebf2170e22161fa5aed3d9025eb8e732e1a3f61903cc400905a5051eb2433"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.849833 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.851695 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" event={"ID":"a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a","Type":"ContainerStarted","Data":"f179cfb5cf0fbae5b57511697979af5dfd96cf5d2732f498e125dd3f026c5406"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.852138 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.856337 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" event={"ID":"f8b1c8b4-b491-44aa-885e-c6215289cb0d","Type":"ContainerStarted","Data":"e7e3331c5baa27cfa1786a6d7d2679d4eeb650e1a4e6ee6fa086d6819397eadf"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.856975 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.859296 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" event={"ID":"3bdcf078-737b-4a4e-9f17-10a55190af5b","Type":"ContainerStarted","Data":"a22724e6948ac4ee29cbc798a50fec49efa456fbd5ef2e0758916c684d341d4f"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.859433 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.868196 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" podStartSLOduration=2.959818664 podStartE2EDuration="52.86815508s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.744126967 +0000 UTC m=+1135.798353437" lastFinishedPulling="2025-10-02 11:11:43.652463403 +0000 UTC m=+1185.706689853" observedRunningTime="2025-10-02 11:11:44.857344821 +0000 UTC m=+1186.911571281" watchObservedRunningTime="2025-10-02 11:11:44.86815508 +0000 UTC m=+1186.922381550" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.868289 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" event={"ID":"4905929c-80af-4de9-b7c7-4ba9f3b938ad","Type":"ContainerStarted","Data":"2270adc9a820e2c757656134f7fa6d543b28b49060104e9a0d0f9a95f92cea02"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.868544 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.870889 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" event={"ID":"2ed3e86e-2313-4efa-987a-d8209ce59d73","Type":"ContainerStarted","Data":"203e7fdc0f8a9b2ff8122c60b505441a01702df70fd8d04cb85f760208f4c439"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.871301 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.873923 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" event={"ID":"787aba9c-ec9d-4d98-bae0-77048675a98b","Type":"ContainerStarted","Data":"99befad68dd7fcc854be7220239d37e17d654be646a24fdef9e6802de06a1439"} Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.874336 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.883276 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" podStartSLOduration=2.854208891 podStartE2EDuration="52.883256704s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.599818909 +0000 UTC m=+1135.654045359" lastFinishedPulling="2025-10-02 11:11:43.628866692 +0000 UTC m=+1185.683093172" observedRunningTime="2025-10-02 11:11:44.882098573 +0000 UTC m=+1186.936325033" watchObservedRunningTime="2025-10-02 11:11:44.883256704 +0000 UTC m=+1186.937483154" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.899802 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" podStartSLOduration=2.840923535 podStartE2EDuration="52.899780226s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.613146635 +0000 UTC m=+1135.667373085" lastFinishedPulling="2025-10-02 11:11:43.672003306 +0000 UTC m=+1185.726229776" observedRunningTime="2025-10-02 11:11:44.896420816 +0000 UTC m=+1186.950647276" watchObservedRunningTime="2025-10-02 11:11:44.899780226 +0000 UTC m=+1186.954006686" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.930422 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" podStartSLOduration=3.079910024 podStartE2EDuration="52.930405434s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.778325871 +0000 UTC m=+1135.832552321" lastFinishedPulling="2025-10-02 11:11:43.628821281 +0000 UTC m=+1185.683047731" observedRunningTime="2025-10-02 11:11:44.91604012 +0000 UTC m=+1186.970266600" watchObservedRunningTime="2025-10-02 11:11:44.930405434 +0000 UTC m=+1186.984631894" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.931677 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" podStartSLOduration=3.095676475 podStartE2EDuration="52.931669158s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.75099148 +0000 UTC m=+1135.805217930" lastFinishedPulling="2025-10-02 11:11:43.586984153 +0000 UTC m=+1185.641210613" observedRunningTime="2025-10-02 11:11:44.929202782 +0000 UTC m=+1186.983429242" watchObservedRunningTime="2025-10-02 11:11:44.931669158 +0000 UTC m=+1186.985895608" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.988933 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" podStartSLOduration=3.059914719 podStartE2EDuration="52.988902228s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.742644017 +0000 UTC m=+1135.796870467" lastFinishedPulling="2025-10-02 11:11:43.671631516 +0000 UTC m=+1185.725857976" observedRunningTime="2025-10-02 11:11:44.972390457 +0000 UTC m=+1187.026616907" watchObservedRunningTime="2025-10-02 11:11:44.988902228 +0000 UTC m=+1187.043128698" Oct 02 11:11:44 crc kubenswrapper[4751]: I1002 11:11:44.990741 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" podStartSLOduration=2.5876941860000002 podStartE2EDuration="52.990718037s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.224209378 +0000 UTC m=+1135.278435828" lastFinishedPulling="2025-10-02 11:11:43.627233189 +0000 UTC m=+1185.681459679" observedRunningTime="2025-10-02 11:11:44.947707097 +0000 UTC m=+1187.001933547" watchObservedRunningTime="2025-10-02 11:11:44.990718037 +0000 UTC m=+1187.044944487" Oct 02 11:11:45 crc kubenswrapper[4751]: I1002 11:11:45.018203 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" podStartSLOduration=3.917588556 podStartE2EDuration="53.018181351s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:54.569435628 +0000 UTC m=+1136.623662078" lastFinishedPulling="2025-10-02 11:11:43.670028423 +0000 UTC m=+1185.724254873" observedRunningTime="2025-10-02 11:11:45.017257686 +0000 UTC m=+1187.071484136" watchObservedRunningTime="2025-10-02 11:11:45.018181351 +0000 UTC m=+1187.072407801" Oct 02 11:11:45 crc kubenswrapper[4751]: I1002 11:11:45.037892 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" podStartSLOduration=2.659252319 podStartE2EDuration="53.037876827s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.292879204 +0000 UTC m=+1135.347105654" lastFinishedPulling="2025-10-02 11:11:43.671503712 +0000 UTC m=+1185.725730162" observedRunningTime="2025-10-02 11:11:45.029871173 +0000 UTC m=+1187.084097623" watchObservedRunningTime="2025-10-02 11:11:45.037876827 +0000 UTC m=+1187.092103277" Oct 02 11:11:45 crc kubenswrapper[4751]: I1002 11:11:45.048692 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" podStartSLOduration=2.983431375 podStartE2EDuration="53.048674536s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.605738357 +0000 UTC m=+1135.659964807" lastFinishedPulling="2025-10-02 11:11:43.670981518 +0000 UTC m=+1185.725207968" observedRunningTime="2025-10-02 11:11:45.047919226 +0000 UTC m=+1187.102145676" watchObservedRunningTime="2025-10-02 11:11:45.048674536 +0000 UTC m=+1187.102900986" Oct 02 11:11:46 crc kubenswrapper[4751]: E1002 11:11:46.553194 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podUID="64d6192e-f762-40ea-8d3f-9a2f8f27144b" Oct 02 11:11:48 crc kubenswrapper[4751]: E1002 11:11:48.551469 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podUID="738b4521-cee5-4fa7-95df-3ac0ea68b26f" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.425083 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kglk4" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.475950 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-clwc9" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.637521 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-4rhhf" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.673606 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-6lcbb" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.715832 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-jk8xr" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.738497 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-kqkzd" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.803905 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-vq88r" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.811698 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-hcnpc" Oct 02 11:11:52 crc kubenswrapper[4751]: I1002 11:11:52.834881 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-vrndj" Oct 02 11:11:54 crc kubenswrapper[4751]: I1002 11:11:54.352281 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-h7qc4" Oct 02 11:11:59 crc kubenswrapper[4751]: I1002 11:11:59.988473 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" event={"ID":"77ffa158-937f-44d2-b65a-9f448a064288","Type":"ContainerStarted","Data":"829af4b7bdec686298b4862f542dac12a4ae862a5551ea0230eec57aa99443e1"} Oct 02 11:11:59 crc kubenswrapper[4751]: I1002 11:11:59.989327 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:11:59 crc kubenswrapper[4751]: I1002 11:11:59.990910 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" event={"ID":"42842b0c-bdf5-4b32-8c6d-f635c333c1b7","Type":"ContainerStarted","Data":"3b847b0e601177ca70124d4599433c95427e289787b2b682013bc738311dadb5"} Oct 02 11:11:59 crc kubenswrapper[4751]: I1002 11:11:59.993932 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" event={"ID":"64d6192e-f762-40ea-8d3f-9a2f8f27144b","Type":"ContainerStarted","Data":"cf85bfdc6d6a74af7d74f0159e103b779db71c1716ad092e1b2fb066dae1c946"} Oct 02 11:12:00 crc kubenswrapper[4751]: I1002 11:12:00.015233 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" podStartSLOduration=3.185181179 podStartE2EDuration="1m8.015208649s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.946409434 +0000 UTC m=+1136.000635884" lastFinishedPulling="2025-10-02 11:11:58.776436904 +0000 UTC m=+1200.830663354" observedRunningTime="2025-10-02 11:12:00.007322618 +0000 UTC m=+1202.061549078" watchObservedRunningTime="2025-10-02 11:12:00.015208649 +0000 UTC m=+1202.069435119" Oct 02 11:12:01 crc kubenswrapper[4751]: I1002 11:12:01.028645 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" podStartSLOduration=3.357537476 podStartE2EDuration="1m9.02862948s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.887010316 +0000 UTC m=+1135.941236766" lastFinishedPulling="2025-10-02 11:11:59.5581023 +0000 UTC m=+1201.612328770" observedRunningTime="2025-10-02 11:12:01.025793244 +0000 UTC m=+1203.080019694" watchObservedRunningTime="2025-10-02 11:12:01.02862948 +0000 UTC m=+1203.082855930" Oct 02 11:12:01 crc kubenswrapper[4751]: I1002 11:12:01.056903 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" podStartSLOduration=3.450268155 podStartE2EDuration="1m9.056883525s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.953706369 +0000 UTC m=+1136.007932819" lastFinishedPulling="2025-10-02 11:11:59.560321739 +0000 UTC m=+1201.614548189" observedRunningTime="2025-10-02 11:12:01.05218495 +0000 UTC m=+1203.106411410" watchObservedRunningTime="2025-10-02 11:12:01.056883525 +0000 UTC m=+1203.111109975" Oct 02 11:12:02 crc kubenswrapper[4751]: I1002 11:12:02.939106 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:12:02 crc kubenswrapper[4751]: I1002 11:12:02.961159 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:12:03 crc kubenswrapper[4751]: I1002 11:12:03.010210 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-q2bvx" Oct 02 11:12:05 crc kubenswrapper[4751]: I1002 11:12:05.033842 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" event={"ID":"738b4521-cee5-4fa7-95df-3ac0ea68b26f","Type":"ContainerStarted","Data":"b185ed1a906637d45e1be6ad007a001476da1337d731e9dac9ed875f404fb9f4"} Oct 02 11:12:05 crc kubenswrapper[4751]: I1002 11:12:05.034457 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:12:05 crc kubenswrapper[4751]: I1002 11:12:05.053323 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" podStartSLOduration=2.496435298 podStartE2EDuration="1m13.053305817s" podCreationTimestamp="2025-10-02 11:10:52 +0000 UTC" firstStartedPulling="2025-10-02 11:10:53.946428555 +0000 UTC m=+1136.000655005" lastFinishedPulling="2025-10-02 11:12:04.503299074 +0000 UTC m=+1206.557525524" observedRunningTime="2025-10-02 11:12:05.051058357 +0000 UTC m=+1207.105284807" watchObservedRunningTime="2025-10-02 11:12:05.053305817 +0000 UTC m=+1207.107532267" Oct 02 11:12:12 crc kubenswrapper[4751]: I1002 11:12:12.942270 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-flblb" Oct 02 11:12:12 crc kubenswrapper[4751]: I1002 11:12:12.957092 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-j85jt" Oct 02 11:12:12 crc kubenswrapper[4751]: I1002 11:12:12.965946 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-pzqdp" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.162152 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.164985 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.173255 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.173485 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.173606 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.173715 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gznjq" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.175181 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.186708 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.186750 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7f2n\" (UniqueName: \"kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.218628 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.220090 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.226349 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.227591 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.287824 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.287894 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.287954 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g86d\" (UniqueName: \"kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.288015 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.288053 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7f2n\" (UniqueName: \"kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.289076 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.307880 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7f2n\" (UniqueName: \"kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n\") pod \"dnsmasq-dns-675f4bcbfc-zvgj8\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.388575 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.388863 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g86d\" (UniqueName: \"kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.389005 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.389656 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.389796 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.406069 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g86d\" (UniqueName: \"kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d\") pod \"dnsmasq-dns-78dd6ddcc-l557d\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.489517 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.538370 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:12:32 crc kubenswrapper[4751]: I1002 11:12:32.946217 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:12:33 crc kubenswrapper[4751]: I1002 11:12:33.013534 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:12:33 crc kubenswrapper[4751]: I1002 11:12:33.265123 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" event={"ID":"e84c5301-6de3-4f6f-882b-c81c053ca50b","Type":"ContainerStarted","Data":"bda6ba7bfa91a7376fa1e8fd11f078a2debda3f951034741b0e6869ba48fb02f"} Oct 02 11:12:33 crc kubenswrapper[4751]: I1002 11:12:33.267237 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" event={"ID":"367b1947-35b4-4169-ad99-fcc21cb25fde","Type":"ContainerStarted","Data":"3b34cffb547cacf99f81a2300c85484c870bf4a341346cad04b9cc0148a88e66"} Oct 02 11:12:34 crc kubenswrapper[4751]: I1002 11:12:34.949271 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:12:34 crc kubenswrapper[4751]: I1002 11:12:34.977820 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:12:34 crc kubenswrapper[4751]: I1002 11:12:34.979128 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:34 crc kubenswrapper[4751]: I1002 11:12:34.990462 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.043453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgkl\" (UniqueName: \"kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.043563 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.043618 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.149900 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgkl\" (UniqueName: \"kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.150012 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.150066 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.151221 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.152136 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.186253 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgkl\" (UniqueName: \"kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl\") pod \"dnsmasq-dns-666b6646f7-kjfln\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.234687 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.267422 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.271799 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.280030 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.314711 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.352895 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.352947 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kglz\" (UniqueName: \"kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.353095 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.454623 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kglz\" (UniqueName: \"kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.454712 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.454765 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.455812 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.456364 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.484251 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kglz\" (UniqueName: \"kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz\") pod \"dnsmasq-dns-57d769cc4f-2tdxk\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.607599 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:12:35 crc kubenswrapper[4751]: I1002 11:12:35.929502 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:12:35 crc kubenswrapper[4751]: W1002 11:12:35.941322 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53608afd_2a0e_4f7f_ae34_ab1c2848b530.slice/crio-52c1678dfddf6aacec5523a9aab91f78950db6a225bcf798bc6767ce491df44f WatchSource:0}: Error finding container 52c1678dfddf6aacec5523a9aab91f78950db6a225bcf798bc6767ce491df44f: Status 404 returned error can't find the container with id 52c1678dfddf6aacec5523a9aab91f78950db6a225bcf798bc6767ce491df44f Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.084194 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:12:36 crc kubenswrapper[4751]: W1002 11:12:36.087143 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9991d5e7_fcb2_4535_b9c8_d327ef696edc.slice/crio-7fd1b4cf94bcb4b84f792e856647c845b948f68f064e7ed0f966eab35c0615df WatchSource:0}: Error finding container 7fd1b4cf94bcb4b84f792e856647c845b948f68f064e7ed0f966eab35c0615df: Status 404 returned error can't find the container with id 7fd1b4cf94bcb4b84f792e856647c845b948f68f064e7ed0f966eab35c0615df Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.129262 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.130458 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.132573 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.132761 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.133142 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.133148 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.133474 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4q2c4" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.135214 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.139088 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.142135 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170276 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170455 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170507 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170532 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170567 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqr24\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170640 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170691 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170722 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170752 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.170784 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.271858 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.271911 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.271927 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.271953 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.271970 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272010 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqr24\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272037 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272069 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272091 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272112 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272137 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.272983 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.273904 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.273965 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.275159 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.277377 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.278149 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.278943 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.278989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.279219 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.292239 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.293726 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqr24\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.299209 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.315958 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" event={"ID":"53608afd-2a0e-4f7f-ae34-ab1c2848b530","Type":"ContainerStarted","Data":"52c1678dfddf6aacec5523a9aab91f78950db6a225bcf798bc6767ce491df44f"} Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.317435 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" event={"ID":"9991d5e7-fcb2-4535-b9c8-d327ef696edc","Type":"ContainerStarted","Data":"7fd1b4cf94bcb4b84f792e856647c845b948f68f064e7ed0f966eab35c0615df"} Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.402806 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.417333 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.417356 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.419256 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zv4vl" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.420974 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.421144 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.421183 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.421255 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.421316 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.421365 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.464409 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.475979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476054 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476085 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476117 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476218 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476306 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476335 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9f8\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476367 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476399 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476444 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.476474 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578092 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578307 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578447 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578486 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578726 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578754 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578779 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578946 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.578992 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9f8\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.579539 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.579628 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.580312 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.580316 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.580356 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.580496 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.580521 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.581068 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.603408 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.603449 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.603557 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.603640 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.609825 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9f8\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.633393 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.740692 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:12:36 crc kubenswrapper[4751]: I1002 11:12:36.946577 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:12:36 crc kubenswrapper[4751]: W1002 11:12:36.953862 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod433c0079_6192_4515_8c6d_5f035612f44b.slice/crio-ec900eeada206aa078b73b92d7513119d67c69fccfcd33f677eaccdf7351f171 WatchSource:0}: Error finding container ec900eeada206aa078b73b92d7513119d67c69fccfcd33f677eaccdf7351f171: Status 404 returned error can't find the container with id ec900eeada206aa078b73b92d7513119d67c69fccfcd33f677eaccdf7351f171 Oct 02 11:12:37 crc kubenswrapper[4751]: I1002 11:12:37.210289 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:12:37 crc kubenswrapper[4751]: W1002 11:12:37.229216 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod341995a5_cf9c_42ce_9afb_b7b50b0c6c4b.slice/crio-b39c67bf08faca1f6ff89d0963058cd6fe8375995163a674a7d9dcb3abf282e9 WatchSource:0}: Error finding container b39c67bf08faca1f6ff89d0963058cd6fe8375995163a674a7d9dcb3abf282e9: Status 404 returned error can't find the container with id b39c67bf08faca1f6ff89d0963058cd6fe8375995163a674a7d9dcb3abf282e9 Oct 02 11:12:37 crc kubenswrapper[4751]: I1002 11:12:37.326030 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerStarted","Data":"b39c67bf08faca1f6ff89d0963058cd6fe8375995163a674a7d9dcb3abf282e9"} Oct 02 11:12:37 crc kubenswrapper[4751]: I1002 11:12:37.328055 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerStarted","Data":"ec900eeada206aa078b73b92d7513119d67c69fccfcd33f677eaccdf7351f171"} Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.127047 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.128465 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.134242 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.134317 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lwxsq" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.134484 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.134994 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.135209 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.138031 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.147532 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213369 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-default\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213623 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213661 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kolla-config\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213679 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-secrets\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213701 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213721 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcs6f\" (UniqueName: \"kubernetes.io/projected/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kube-api-access-fcs6f\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213742 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213776 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.213798 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315239 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315367 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315428 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-default\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315458 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315506 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kolla-config\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315531 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-secrets\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315561 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.315588 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcs6f\" (UniqueName: \"kubernetes.io/projected/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kube-api-access-fcs6f\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.316313 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.316742 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.317146 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-config-data-default\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.317477 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.319870 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kolla-config\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.323410 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-secrets\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.330166 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.333111 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcs6f\" (UniqueName: \"kubernetes.io/projected/6bd070ec-aa1f-4991-a239-df38cfff5fcf-kube-api-access-fcs6f\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.336708 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd070ec-aa1f-4991-a239-df38cfff5fcf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.339040 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"6bd070ec-aa1f-4991-a239-df38cfff5fcf\") " pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.448665 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.876227 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.878571 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.886396 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-h9js5" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.886575 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.886804 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.887229 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 11:12:38 crc kubenswrapper[4751]: I1002 11:12:38.894568 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025249 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025433 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025496 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vljwd\" (UniqueName: \"kubernetes.io/projected/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kube-api-access-vljwd\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025520 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025563 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025603 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025691 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.025727 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127212 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127271 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127308 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127326 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127350 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127420 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127445 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vljwd\" (UniqueName: \"kubernetes.io/projected/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kube-api-access-vljwd\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127460 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127710 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.127983 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.128763 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.130503 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.131753 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.132749 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.132994 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.133163 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.146011 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vljwd\" (UniqueName: \"kubernetes.io/projected/fa2f2972-7b83-44e7-8ba5-37fb04475f5c-kube-api-access-vljwd\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.154063 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"fa2f2972-7b83-44e7-8ba5-37fb04475f5c\") " pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.213772 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.401453 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.402860 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.404989 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.405598 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cp4sn" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.405866 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.409091 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.536075 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.536129 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-config-data\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.536162 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-kolla-config\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.536281 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.536309 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h58x9\" (UniqueName: \"kubernetes.io/projected/2b5f725c-7545-484d-a3c4-37e69cf21a89-kube-api-access-h58x9\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.638152 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.638252 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-config-data\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.638473 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-kolla-config\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.638536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.638556 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h58x9\" (UniqueName: \"kubernetes.io/projected/2b5f725c-7545-484d-a3c4-37e69cf21a89-kube-api-access-h58x9\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.639004 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-config-data\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.640379 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2b5f725c-7545-484d-a3c4-37e69cf21a89-kolla-config\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.643406 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.657052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b5f725c-7545-484d-a3c4-37e69cf21a89-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.666367 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h58x9\" (UniqueName: \"kubernetes.io/projected/2b5f725c-7545-484d-a3c4-37e69cf21a89-kube-api-access-h58x9\") pod \"memcached-0\" (UID: \"2b5f725c-7545-484d-a3c4-37e69cf21a89\") " pod="openstack/memcached-0" Oct 02 11:12:39 crc kubenswrapper[4751]: I1002 11:12:39.738959 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.336181 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.337346 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.339589 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-jpqdf" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.349832 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.366684 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghp4\" (UniqueName: \"kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4\") pod \"kube-state-metrics-0\" (UID: \"82876497-5a0e-4912-8e2f-e5e5a43f5c95\") " pod="openstack/kube-state-metrics-0" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.468038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghp4\" (UniqueName: \"kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4\") pod \"kube-state-metrics-0\" (UID: \"82876497-5a0e-4912-8e2f-e5e5a43f5c95\") " pod="openstack/kube-state-metrics-0" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.487311 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghp4\" (UniqueName: \"kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4\") pod \"kube-state-metrics-0\" (UID: \"82876497-5a0e-4912-8e2f-e5e5a43f5c95\") " pod="openstack/kube-state-metrics-0" Oct 02 11:12:41 crc kubenswrapper[4751]: I1002 11:12:41.671467 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.608978 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tqqkd"] Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.610595 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.612647 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-sq2gf" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.612771 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.612794 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618336 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618414 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868943fd-a908-4e94-bef5-737ba8ebc1da-scripts\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618504 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwd5r\" (UniqueName: \"kubernetes.io/projected/868943fd-a908-4e94-bef5-737ba8ebc1da-kube-api-access-lwd5r\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618543 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-combined-ca-bundle\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618569 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-ovn-controller-tls-certs\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618704 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-log-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.618747 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.626787 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tqqkd"] Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.678228 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-t54r5"] Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.679779 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.683460 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t54r5"] Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720585 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-combined-ca-bundle\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720655 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-ovn-controller-tls-certs\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720736 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a41c88c-74de-4518-8313-441084f2cf0f-scripts\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720770 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-lib\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720828 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-run\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720883 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-log-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720936 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-etc-ovs\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.720997 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.721029 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-log\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.721062 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26sc\" (UniqueName: \"kubernetes.io/projected/7a41c88c-74de-4518-8313-441084f2cf0f-kube-api-access-c26sc\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.721097 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.721132 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868943fd-a908-4e94-bef5-737ba8ebc1da-scripts\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.721223 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwd5r\" (UniqueName: \"kubernetes.io/projected/868943fd-a908-4e94-bef5-737ba8ebc1da-kube-api-access-lwd5r\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.722206 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-log-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.722394 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.723279 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/868943fd-a908-4e94-bef5-737ba8ebc1da-var-run-ovn\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.724639 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/868943fd-a908-4e94-bef5-737ba8ebc1da-scripts\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.727302 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-ovn-controller-tls-certs\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.727456 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/868943fd-a908-4e94-bef5-737ba8ebc1da-combined-ca-bundle\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.736242 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwd5r\" (UniqueName: \"kubernetes.io/projected/868943fd-a908-4e94-bef5-737ba8ebc1da-kube-api-access-lwd5r\") pod \"ovn-controller-tqqkd\" (UID: \"868943fd-a908-4e94-bef5-737ba8ebc1da\") " pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821645 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a41c88c-74de-4518-8313-441084f2cf0f-scripts\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821684 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-lib\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821702 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-run\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821744 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-etc-ovs\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821765 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-log\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821784 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26sc\" (UniqueName: \"kubernetes.io/projected/7a41c88c-74de-4518-8313-441084f2cf0f-kube-api-access-c26sc\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.821958 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-run\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.822030 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-etc-ovs\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.822032 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-lib\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.822149 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7a41c88c-74de-4518-8313-441084f2cf0f-var-log\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.824568 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a41c88c-74de-4518-8313-441084f2cf0f-scripts\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.837311 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26sc\" (UniqueName: \"kubernetes.io/projected/7a41c88c-74de-4518-8313-441084f2cf0f-kube-api-access-c26sc\") pod \"ovn-controller-ovs-t54r5\" (UID: \"7a41c88c-74de-4518-8313-441084f2cf0f\") " pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:44 crc kubenswrapper[4751]: I1002 11:12:44.930358 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.003321 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.101329 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.102871 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.106447 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.106512 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.106623 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.106737 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.108893 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-g6ffj" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.115189 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.226971 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227027 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/539d6784-c995-47a3-9a80-53223cd9da09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227068 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-config\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227093 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227113 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227182 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227248 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.227380 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg4mf\" (UniqueName: \"kubernetes.io/projected/539d6784-c995-47a3-9a80-53223cd9da09-kube-api-access-qg4mf\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328769 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328839 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/539d6784-c995-47a3-9a80-53223cd9da09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328867 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-config\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328908 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.328954 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.329019 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.329062 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg4mf\" (UniqueName: \"kubernetes.io/projected/539d6784-c995-47a3-9a80-53223cd9da09-kube-api-access-qg4mf\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.329315 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.329767 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-config\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.329884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/539d6784-c995-47a3-9a80-53223cd9da09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.330885 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/539d6784-c995-47a3-9a80-53223cd9da09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.332855 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.332856 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.336496 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/539d6784-c995-47a3-9a80-53223cd9da09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.348997 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg4mf\" (UniqueName: \"kubernetes.io/projected/539d6784-c995-47a3-9a80-53223cd9da09-kube-api-access-qg4mf\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.356285 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"539d6784-c995-47a3-9a80-53223cd9da09\") " pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:45 crc kubenswrapper[4751]: I1002 11:12:45.435817 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.023422 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.026098 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.027821 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.027840 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-wdxtz" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.029037 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.029085 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.034296 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192316 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192408 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-config\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192531 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192602 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s868k\" (UniqueName: \"kubernetes.io/projected/6b3aa529-c99f-464f-85a0-70711c59f08f-kube-api-access-s868k\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192650 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192743 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192788 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.192927 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294506 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294583 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294676 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294767 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294799 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-config\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294863 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294894 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s868k\" (UniqueName: \"kubernetes.io/projected/6b3aa529-c99f-464f-85a0-70711c59f08f-kube-api-access-s868k\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.294921 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.295456 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.295564 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.296605 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-config\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.296994 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b3aa529-c99f-464f-85a0-70711c59f08f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.309049 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.311201 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.325752 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b3aa529-c99f-464f-85a0-70711c59f08f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.335581 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s868k\" (UniqueName: \"kubernetes.io/projected/6b3aa529-c99f-464f-85a0-70711c59f08f-kube-api-access-s868k\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.363614 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6b3aa529-c99f-464f-85a0-70711c59f08f\") " pod="openstack/ovsdbserver-sb-0" Oct 02 11:12:49 crc kubenswrapper[4751]: I1002 11:12:49.660670 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.509380 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.510130 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tx9f8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(341995a5-cf9c-42ce-9afb-b7b50b0c6c4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.511355 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.635456 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.635708 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s7f2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-zvgj8_openstack(367b1947-35b4-4169-ad99-fcc21cb25fde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.636894 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" podUID="367b1947-35b4-4169-ad99-fcc21cb25fde" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.781099 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.992244 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.992704 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6g86d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-l557d_openstack(e84c5301-6de3-4f6f-882b-c81c053ca50b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:05 crc kubenswrapper[4751]: E1002 11:13:05.993915 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" podUID="e84c5301-6de3-4f6f-882b-c81c053ca50b" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.105494 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.105616 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-srgkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-kjfln_openstack(53608afd-2a0e-4f7f-ae34-ab1c2848b530): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.106951 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.230506 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: W1002 11:13:06.231761 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa2f2972_7b83_44e7_8ba5_37fb04475f5c.slice/crio-77095e26063497caf472fc82c7569db68ebbbb35f91355354a89dd692fde198d WatchSource:0}: Error finding container 77095e26063497caf472fc82c7569db68ebbbb35f91355354a89dd692fde198d: Status 404 returned error can't find the container with id 77095e26063497caf472fc82c7569db68ebbbb35f91355354a89dd692fde198d Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.303778 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.303928 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kglz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-2tdxk_openstack(9991d5e7-fcb2-4535-b9c8-d327ef696edc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.305662 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.396705 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.404656 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tqqkd"] Oct 02 11:13:06 crc kubenswrapper[4751]: W1002 11:13:06.416993 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod868943fd_a908_4e94_bef5_737ba8ebc1da.slice/crio-c61cfea514e7035d587357d39ff67db7329f3c1c6d31babe9907f440055b0886 WatchSource:0}: Error finding container c61cfea514e7035d587357d39ff67db7329f3c1c6d31babe9907f440055b0886: Status 404 returned error can't find the container with id c61cfea514e7035d587357d39ff67db7329f3c1c6d31babe9907f440055b0886 Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.485486 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: W1002 11:13:06.491332 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b3aa529_c99f_464f_85a0_70711c59f08f.slice/crio-340ed65d0b8c4d13522c719312ef9b9c7bc6cc2a711d42c92d3f3136cac1c726 WatchSource:0}: Error finding container 340ed65d0b8c4d13522c719312ef9b9c7bc6cc2a711d42c92d3f3136cac1c726: Status 404 returned error can't find the container with id 340ed65d0b8c4d13522c719312ef9b9c7bc6cc2a711d42c92d3f3136cac1c726 Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.570340 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fa2f2972-7b83-44e7-8ba5-37fb04475f5c","Type":"ContainerStarted","Data":"77095e26063497caf472fc82c7569db68ebbbb35f91355354a89dd692fde198d"} Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.573870 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tqqkd" event={"ID":"868943fd-a908-4e94-bef5-737ba8ebc1da","Type":"ContainerStarted","Data":"c61cfea514e7035d587357d39ff67db7329f3c1c6d31babe9907f440055b0886"} Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.575499 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6b3aa529-c99f-464f-85a0-70711c59f08f","Type":"ContainerStarted","Data":"340ed65d0b8c4d13522c719312ef9b9c7bc6cc2a711d42c92d3f3136cac1c726"} Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.577345 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6bd070ec-aa1f-4991-a239-df38cfff5fcf","Type":"ContainerStarted","Data":"3d3ed3a01c3ad41565d806c03f2558f9c6f114d04077a5b9bdffeeb7d6926f99"} Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.579648 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" Oct 02 11:13:06 crc kubenswrapper[4751]: E1002 11:13:06.579750 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.582193 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.595317 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.681257 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.969628 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:13:06 crc kubenswrapper[4751]: I1002 11:13:06.975078 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088062 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc\") pod \"e84c5301-6de3-4f6f-882b-c81c053ca50b\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088424 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config\") pod \"367b1947-35b4-4169-ad99-fcc21cb25fde\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088521 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config\") pod \"e84c5301-6de3-4f6f-882b-c81c053ca50b\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088560 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g86d\" (UniqueName: \"kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d\") pod \"e84c5301-6de3-4f6f-882b-c81c053ca50b\" (UID: \"e84c5301-6de3-4f6f-882b-c81c053ca50b\") " Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088586 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7f2n\" (UniqueName: \"kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n\") pod \"367b1947-35b4-4169-ad99-fcc21cb25fde\" (UID: \"367b1947-35b4-4169-ad99-fcc21cb25fde\") " Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.088731 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e84c5301-6de3-4f6f-882b-c81c053ca50b" (UID: "e84c5301-6de3-4f6f-882b-c81c053ca50b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.089004 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.089340 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config" (OuterVolumeSpecName: "config") pod "367b1947-35b4-4169-ad99-fcc21cb25fde" (UID: "367b1947-35b4-4169-ad99-fcc21cb25fde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.089521 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config" (OuterVolumeSpecName: "config") pod "e84c5301-6de3-4f6f-882b-c81c053ca50b" (UID: "e84c5301-6de3-4f6f-882b-c81c053ca50b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.094338 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d" (OuterVolumeSpecName: "kube-api-access-6g86d") pod "e84c5301-6de3-4f6f-882b-c81c053ca50b" (UID: "e84c5301-6de3-4f6f-882b-c81c053ca50b"). InnerVolumeSpecName "kube-api-access-6g86d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.094663 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n" (OuterVolumeSpecName: "kube-api-access-s7f2n") pod "367b1947-35b4-4169-ad99-fcc21cb25fde" (UID: "367b1947-35b4-4169-ad99-fcc21cb25fde"). InnerVolumeSpecName "kube-api-access-s7f2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.191139 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e84c5301-6de3-4f6f-882b-c81c053ca50b-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.191206 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g86d\" (UniqueName: \"kubernetes.io/projected/e84c5301-6de3-4f6f-882b-c81c053ca50b-kube-api-access-6g86d\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.191226 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7f2n\" (UniqueName: \"kubernetes.io/projected/367b1947-35b4-4169-ad99-fcc21cb25fde-kube-api-access-s7f2n\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.191245 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/367b1947-35b4-4169-ad99-fcc21cb25fde-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.360559 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t54r5"] Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.585586 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t54r5" event={"ID":"7a41c88c-74de-4518-8313-441084f2cf0f","Type":"ContainerStarted","Data":"988bc3b2647bf90722c3cd9f48668ba019afc7a4f7b192c7d7f7bec925129b01"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.587988 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82876497-5a0e-4912-8e2f-e5e5a43f5c95","Type":"ContainerStarted","Data":"b3000dd8e26dadbb4f0f0ef9d870effd8690755ae435ee3c3f94414e7e7437b2"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.589205 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2b5f725c-7545-484d-a3c4-37e69cf21a89","Type":"ContainerStarted","Data":"4e8469c6879b9a58fb641a3180b327fa61b3a2545604136293d432fb1aad8a69"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.590624 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" event={"ID":"e84c5301-6de3-4f6f-882b-c81c053ca50b","Type":"ContainerDied","Data":"bda6ba7bfa91a7376fa1e8fd11f078a2debda3f951034741b0e6869ba48fb02f"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.590699 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-l557d" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.592466 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"539d6784-c995-47a3-9a80-53223cd9da09","Type":"ContainerStarted","Data":"e5cd61c75663abddf9c8f8da32d0852b1fbf08cf3d93088827c206d90fa0a30c"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.599052 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.599058 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-zvgj8" event={"ID":"367b1947-35b4-4169-ad99-fcc21cb25fde","Type":"ContainerDied","Data":"3b34cffb547cacf99f81a2300c85484c870bf4a341346cad04b9cc0148a88e66"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.601195 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerStarted","Data":"8d0714368474f6ee386d468f8266d4e2d0c5105a27266788a8998b264c68bb24"} Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.651849 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:13:07 crc kubenswrapper[4751]: E1002 11:13:07.661321 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode84c5301_6de3_4f6f_882b_c81c053ca50b.slice/crio-bda6ba7bfa91a7376fa1e8fd11f078a2debda3f951034741b0e6869ba48fb02f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod367b1947_35b4_4169_ad99_fcc21cb25fde.slice\": RecentStats: unable to find data in memory cache]" Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.663314 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-zvgj8"] Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.689414 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:13:07 crc kubenswrapper[4751]: I1002 11:13:07.694735 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-l557d"] Oct 02 11:13:09 crc kubenswrapper[4751]: I1002 11:13:09.561197 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="367b1947-35b4-4169-ad99-fcc21cb25fde" path="/var/lib/kubelet/pods/367b1947-35b4-4169-ad99-fcc21cb25fde/volumes" Oct 02 11:13:09 crc kubenswrapper[4751]: I1002 11:13:09.562262 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e84c5301-6de3-4f6f-882b-c81c053ca50b" path="/var/lib/kubelet/pods/e84c5301-6de3-4f6f-882b-c81c053ca50b/volumes" Oct 02 11:13:21 crc kubenswrapper[4751]: I1002 11:13:21.738389 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2b5f725c-7545-484d-a3c4-37e69cf21a89","Type":"ContainerStarted","Data":"2977f4b6e8754c36342971bf5fa8510692072175e07819bd98c658c337b620e0"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.749598 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6bd070ec-aa1f-4991-a239-df38cfff5fcf","Type":"ContainerStarted","Data":"178ea8c0ed18701c2772622ccb984118819d36cec4a35073620b9e266ce05e63"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.751952 4751 generic.go:334] "Generic (PLEG): container finished" podID="7a41c88c-74de-4518-8313-441084f2cf0f" containerID="0ef6b952e0703dd7057ed5e06100ce347f759db8c36e9496e872ac7006abb786" exitCode=0 Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.752031 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t54r5" event={"ID":"7a41c88c-74de-4518-8313-441084f2cf0f","Type":"ContainerDied","Data":"0ef6b952e0703dd7057ed5e06100ce347f759db8c36e9496e872ac7006abb786"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.753908 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fa2f2972-7b83-44e7-8ba5-37fb04475f5c","Type":"ContainerStarted","Data":"30ff247f02bb4f624c9c0248804548ee427f55915812bac78daf34811a60d8d6"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.756104 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"539d6784-c995-47a3-9a80-53223cd9da09","Type":"ContainerStarted","Data":"a04a4fdf719d909c68f7a60a4609c833189e240aec2d6c3d3966d62804a46443"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.757770 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerStarted","Data":"3432c82b9e0622ec5f5ece1b81a44963f17f80a50ec2661d0b17a2e1ba4c5a3a"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.759380 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tqqkd" event={"ID":"868943fd-a908-4e94-bef5-737ba8ebc1da","Type":"ContainerStarted","Data":"359727bbcff84bb21c5892d913bfb031e6ee343596f9e8ca7ccb092ab3bbaf36"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.759468 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tqqkd" Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.761090 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6b3aa529-c99f-464f-85a0-70711c59f08f","Type":"ContainerStarted","Data":"2b8d4a3b6a4a1187273f1c36d7bce6cbc494811fc1a562c40b979b1774882e9f"} Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.761147 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.826211 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tqqkd" podStartSLOduration=25.210933086 podStartE2EDuration="38.826191036s" podCreationTimestamp="2025-10-02 11:12:44 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.41927198 +0000 UTC m=+1268.473498430" lastFinishedPulling="2025-10-02 11:13:20.03452993 +0000 UTC m=+1282.088756380" observedRunningTime="2025-10-02 11:13:22.824696316 +0000 UTC m=+1284.878922796" watchObservedRunningTime="2025-10-02 11:13:22.826191036 +0000 UTC m=+1284.880417486" Oct 02 11:13:22 crc kubenswrapper[4751]: I1002 11:13:22.866995 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=30.930362299 podStartE2EDuration="43.866977557s" podCreationTimestamp="2025-10-02 11:12:39 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.591782262 +0000 UTC m=+1268.646008712" lastFinishedPulling="2025-10-02 11:13:19.52839752 +0000 UTC m=+1281.582623970" observedRunningTime="2025-10-02 11:13:22.866235407 +0000 UTC m=+1284.920461867" watchObservedRunningTime="2025-10-02 11:13:22.866977557 +0000 UTC m=+1284.921204007" Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.786894 4751 generic.go:334] "Generic (PLEG): container finished" podID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerID="8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d" exitCode=0 Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.787509 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" event={"ID":"9991d5e7-fcb2-4535-b9c8-d327ef696edc","Type":"ContainerDied","Data":"8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d"} Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.791818 4751 generic.go:334] "Generic (PLEG): container finished" podID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerID="269d0dc10c724dcb1d575817463d7b18b50198d2284329e9affddf90d9119542" exitCode=0 Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.791878 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" event={"ID":"53608afd-2a0e-4f7f-ae34-ab1c2848b530","Type":"ContainerDied","Data":"269d0dc10c724dcb1d575817463d7b18b50198d2284329e9affddf90d9119542"} Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.794871 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t54r5" event={"ID":"7a41c88c-74de-4518-8313-441084f2cf0f","Type":"ContainerStarted","Data":"9e0ddaee91189a7ecfe9ca8765af3fb3cd14c85dc32dab21991dcea884337186"} Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.794917 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t54r5" event={"ID":"7a41c88c-74de-4518-8313-441084f2cf0f","Type":"ContainerStarted","Data":"7ecef5cb8d058200034557ea1eabcf5554f9001696caa34146410ef02bb500c8"} Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.802962 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82876497-5a0e-4912-8e2f-e5e5a43f5c95","Type":"ContainerStarted","Data":"68a263fe7a00e1b5527ce62c4038d2a1285a89c7b1c52046ab5cf03d8c97af0b"} Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.803293 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 11:13:25 crc kubenswrapper[4751]: I1002 11:13:25.825430 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=26.427959925 podStartE2EDuration="44.825408691s" podCreationTimestamp="2025-10-02 11:12:41 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.61564754 +0000 UTC m=+1268.669873990" lastFinishedPulling="2025-10-02 11:13:25.013096306 +0000 UTC m=+1287.067322756" observedRunningTime="2025-10-02 11:13:25.819714108 +0000 UTC m=+1287.873940578" watchObservedRunningTime="2025-10-02 11:13:25.825408691 +0000 UTC m=+1287.879635151" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.824989 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" event={"ID":"9991d5e7-fcb2-4535-b9c8-d327ef696edc","Type":"ContainerStarted","Data":"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0"} Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.825485 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.828080 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" event={"ID":"53608afd-2a0e-4f7f-ae34-ab1c2848b530","Type":"ContainerStarted","Data":"8e8aa23300300b301d357c5293562758433b97aabbb4f5ccbcba02cf70f27a70"} Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.828401 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.828420 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.828639 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.845595 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" podStartSLOduration=2.921423697 podStartE2EDuration="51.845576422s" podCreationTimestamp="2025-10-02 11:12:35 +0000 UTC" firstStartedPulling="2025-10-02 11:12:36.088969032 +0000 UTC m=+1238.143195482" lastFinishedPulling="2025-10-02 11:13:25.013121737 +0000 UTC m=+1287.067348207" observedRunningTime="2025-10-02 11:13:26.844274387 +0000 UTC m=+1288.898500847" watchObservedRunningTime="2025-10-02 11:13:26.845576422 +0000 UTC m=+1288.899802872" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.867593 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-t54r5" podStartSLOduration=30.516585195 podStartE2EDuration="42.867552949s" podCreationTimestamp="2025-10-02 11:12:44 +0000 UTC" firstStartedPulling="2025-10-02 11:13:07.362451582 +0000 UTC m=+1269.416678032" lastFinishedPulling="2025-10-02 11:13:19.713419326 +0000 UTC m=+1281.767645786" observedRunningTime="2025-10-02 11:13:26.864868588 +0000 UTC m=+1288.919095048" watchObservedRunningTime="2025-10-02 11:13:26.867552949 +0000 UTC m=+1288.921779409" Oct 02 11:13:26 crc kubenswrapper[4751]: I1002 11:13:26.884374 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podStartSLOduration=3.813933224 podStartE2EDuration="52.884353938s" podCreationTimestamp="2025-10-02 11:12:34 +0000 UTC" firstStartedPulling="2025-10-02 11:12:35.944239194 +0000 UTC m=+1237.998465644" lastFinishedPulling="2025-10-02 11:13:25.014659908 +0000 UTC m=+1287.068886358" observedRunningTime="2025-10-02 11:13:26.880927907 +0000 UTC m=+1288.935154377" watchObservedRunningTime="2025-10-02 11:13:26.884353938 +0000 UTC m=+1288.938580388" Oct 02 11:13:29 crc kubenswrapper[4751]: I1002 11:13:29.740457 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.605772 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.606349 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" containerID="cri-o://8e8aa23300300b301d357c5293562758433b97aabbb4f5ccbcba02cf70f27a70" gracePeriod=10 Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.607340 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.637838 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.639135 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.661287 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.688522 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.715716 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.716136 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs78s\" (UniqueName: \"kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.716212 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.818004 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.818216 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs78s\" (UniqueName: \"kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.818255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.819404 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.819441 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.843666 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs78s\" (UniqueName: \"kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s\") pod \"dnsmasq-dns-7cb5889db5-fznz8\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:31 crc kubenswrapper[4751]: I1002 11:13:31.962781 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.753618 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.764355 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.766699 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.767161 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.767570 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.771257 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jrh4f" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.791141 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.833589 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-cache\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.833939 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-lock\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.834073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvs5x\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-kube-api-access-zvs5x\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.834279 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.834443 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.962992 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963225 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-cache\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963275 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-lock\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963346 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvs5x\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-kube-api-access-zvs5x\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963404 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: E1002 11:13:32.963614 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:32 crc kubenswrapper[4751]: E1002 11:13:32.963729 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:32 crc kubenswrapper[4751]: E1002 11:13:32.963889 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:13:33.463866694 +0000 UTC m=+1295.518093184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963882 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-cache\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963893 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.963958 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/b0071ad0-2497-45b8-bddc-9164ee537d18-lock\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.980634 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvs5x\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-kube-api-access-zvs5x\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:32 crc kubenswrapper[4751]: I1002 11:13:32.996201 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:33 crc kubenswrapper[4751]: I1002 11:13:33.470937 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:33 crc kubenswrapper[4751]: E1002 11:13:33.471188 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:33 crc kubenswrapper[4751]: E1002 11:13:33.471227 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:33 crc kubenswrapper[4751]: E1002 11:13:33.471292 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:13:34.471271278 +0000 UTC m=+1296.525497728 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:33 crc kubenswrapper[4751]: I1002 11:13:33.892393 4751 generic.go:334] "Generic (PLEG): container finished" podID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerID="8e8aa23300300b301d357c5293562758433b97aabbb4f5ccbcba02cf70f27a70" exitCode=0 Oct 02 11:13:33 crc kubenswrapper[4751]: I1002 11:13:33.892450 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" event={"ID":"53608afd-2a0e-4f7f-ae34-ab1c2848b530","Type":"ContainerDied","Data":"8e8aa23300300b301d357c5293562758433b97aabbb4f5ccbcba02cf70f27a70"} Oct 02 11:13:34 crc kubenswrapper[4751]: I1002 11:13:34.486914 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:34 crc kubenswrapper[4751]: E1002 11:13:34.487119 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:34 crc kubenswrapper[4751]: E1002 11:13:34.487515 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:34 crc kubenswrapper[4751]: E1002 11:13:34.487578 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:13:36.487558806 +0000 UTC m=+1298.541785256 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:35 crc kubenswrapper[4751]: I1002 11:13:35.315650 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.94:5353: connect: connection refused" Oct 02 11:13:35 crc kubenswrapper[4751]: I1002 11:13:35.609372 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.521582 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:36 crc kubenswrapper[4751]: E1002 11:13:36.521771 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:36 crc kubenswrapper[4751]: E1002 11:13:36.521869 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:36 crc kubenswrapper[4751]: E1002 11:13:36.521917 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:13:40.521901738 +0000 UTC m=+1302.576128188 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.719006 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bkkpx"] Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.720337 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.722924 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.723282 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.723921 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.732447 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bkkpx"] Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.826649 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.826707 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.826784 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.826805 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.827057 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfgq2\" (UniqueName: \"kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.827139 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.827199 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928776 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfgq2\" (UniqueName: \"kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928831 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928851 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928910 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928959 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.928977 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.930344 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.930468 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.930836 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.933688 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.934431 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.934987 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:36 crc kubenswrapper[4751]: I1002 11:13:36.959385 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfgq2\" (UniqueName: \"kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2\") pod \"swift-ring-rebalance-bkkpx\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:37 crc kubenswrapper[4751]: I1002 11:13:37.037193 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:13:39 crc kubenswrapper[4751]: I1002 11:13:39.938104 4751 generic.go:334] "Generic (PLEG): container finished" podID="433c0079-6192-4515-8c6d-5f035612f44b" containerID="8d0714368474f6ee386d468f8266d4e2d0c5105a27266788a8998b264c68bb24" exitCode=0 Oct 02 11:13:39 crc kubenswrapper[4751]: I1002 11:13:39.938208 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerDied","Data":"8d0714368474f6ee386d468f8266d4e2d0c5105a27266788a8998b264c68bb24"} Oct 02 11:13:40 crc kubenswrapper[4751]: I1002 11:13:40.315273 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.94:5353: connect: connection refused" Oct 02 11:13:40 crc kubenswrapper[4751]: I1002 11:13:40.590702 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:40 crc kubenswrapper[4751]: E1002 11:13:40.591221 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:40 crc kubenswrapper[4751]: E1002 11:13:40.591249 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:40 crc kubenswrapper[4751]: E1002 11:13:40.591316 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:13:48.591296861 +0000 UTC m=+1310.645523311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.689194 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.809535 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc\") pod \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.809644 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config\") pod \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.810095 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srgkl\" (UniqueName: \"kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl\") pod \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\" (UID: \"53608afd-2a0e-4f7f-ae34-ab1c2848b530\") " Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.814964 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl" (OuterVolumeSpecName: "kube-api-access-srgkl") pod "53608afd-2a0e-4f7f-ae34-ab1c2848b530" (UID: "53608afd-2a0e-4f7f-ae34-ab1c2848b530"). InnerVolumeSpecName "kube-api-access-srgkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.849243 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config" (OuterVolumeSpecName: "config") pod "53608afd-2a0e-4f7f-ae34-ab1c2848b530" (UID: "53608afd-2a0e-4f7f-ae34-ab1c2848b530"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.858563 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53608afd-2a0e-4f7f-ae34-ab1c2848b530" (UID: "53608afd-2a0e-4f7f-ae34-ab1c2848b530"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.913864 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srgkl\" (UniqueName: \"kubernetes.io/projected/53608afd-2a0e-4f7f-ae34-ab1c2848b530-kube-api-access-srgkl\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.913915 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.913925 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53608afd-2a0e-4f7f-ae34-ab1c2848b530-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.955845 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" event={"ID":"53608afd-2a0e-4f7f-ae34-ab1c2848b530","Type":"ContainerDied","Data":"52c1678dfddf6aacec5523a9aab91f78950db6a225bcf798bc6767ce491df44f"} Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.955921 4751 scope.go:117] "RemoveContainer" containerID="8e8aa23300300b301d357c5293562758433b97aabbb4f5ccbcba02cf70f27a70" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.955960 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kjfln" Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.985680 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:13:41 crc kubenswrapper[4751]: I1002 11:13:41.990549 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kjfln"] Oct 02 11:13:43 crc kubenswrapper[4751]: I1002 11:13:43.566666 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" path="/var/lib/kubelet/pods/53608afd-2a0e-4f7f-ae34-ab1c2848b530/volumes" Oct 02 11:13:47 crc kubenswrapper[4751]: I1002 11:13:47.124319 4751 scope.go:117] "RemoveContainer" containerID="269d0dc10c724dcb1d575817463d7b18b50198d2284329e9affddf90d9119542" Oct 02 11:13:47 crc kubenswrapper[4751]: I1002 11:13:47.503675 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bkkpx"] Oct 02 11:13:47 crc kubenswrapper[4751]: W1002 11:13:47.510445 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30a668e4_d395_4b78_a014_279fd35ed304.slice/crio-9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde WatchSource:0}: Error finding container 9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde: Status 404 returned error can't find the container with id 9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde Oct 02 11:13:47 crc kubenswrapper[4751]: I1002 11:13:47.576236 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:47 crc kubenswrapper[4751]: W1002 11:13:47.580018 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3814782_f0fa_40c9_a5ab_2b6d4fda91fe.slice/crio-868fd492a6de2d2549a923feecd1b493860917f8ec605876c8ff4ac948f2868f WatchSource:0}: Error finding container 868fd492a6de2d2549a923feecd1b493860917f8ec605876c8ff4ac948f2868f: Status 404 returned error can't find the container with id 868fd492a6de2d2549a923feecd1b493860917f8ec605876c8ff4ac948f2868f Oct 02 11:13:48 crc kubenswrapper[4751]: I1002 11:13:48.007942 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerStarted","Data":"868fd492a6de2d2549a923feecd1b493860917f8ec605876c8ff4ac948f2868f"} Oct 02 11:13:48 crc kubenswrapper[4751]: I1002 11:13:48.008999 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bkkpx" event={"ID":"30a668e4-d395-4b78-a014-279fd35ed304","Type":"ContainerStarted","Data":"9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde"} Oct 02 11:13:48 crc kubenswrapper[4751]: I1002 11:13:48.628254 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:13:48 crc kubenswrapper[4751]: E1002 11:13:48.628395 4751 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 11:13:48 crc kubenswrapper[4751]: E1002 11:13:48.628766 4751 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 11:13:48 crc kubenswrapper[4751]: E1002 11:13:48.628808 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift podName:b0071ad0-2497-45b8-bddc-9164ee537d18 nodeName:}" failed. No retries permitted until 2025-10-02 11:14:04.628794477 +0000 UTC m=+1326.683020917 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift") pod "swift-storage-0" (UID: "b0071ad0-2497-45b8-bddc-9164ee537d18") : configmap "swift-ring-files" not found Oct 02 11:13:49 crc kubenswrapper[4751]: I1002 11:13:49.021854 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerStarted","Data":"bc6a75489d3fb857d4be16da3b9fd49d8b5d3dfd494165122057d9571f8923a4"} Oct 02 11:13:50 crc kubenswrapper[4751]: E1002 11:13:50.694032 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Oct 02 11:13:50 crc kubenswrapper[4751]: E1002 11:13:50.695021 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qg4mf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(539d6784-c995-47a3-9a80-53223cd9da09): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:50 crc kubenswrapper[4751]: E1002 11:13:50.696232 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="539d6784-c995-47a3-9a80-53223cd9da09" Oct 02 11:13:51 crc kubenswrapper[4751]: I1002 11:13:51.047001 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerStarted","Data":"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd"} Oct 02 11:13:51 crc kubenswrapper[4751]: E1002 11:13:51.048424 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="539d6784-c995-47a3-9a80-53223cd9da09" Oct 02 11:13:51 crc kubenswrapper[4751]: I1002 11:13:51.436463 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 11:13:51 crc kubenswrapper[4751]: I1002 11:13:51.505891 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.065464 4751 generic.go:334] "Generic (PLEG): container finished" podID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerID="f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd" exitCode=0 Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.065883 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerDied","Data":"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd"} Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.066992 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.067072 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 11:13:52 crc kubenswrapper[4751]: E1002 11:13:52.069799 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="539d6784-c995-47a3-9a80-53223cd9da09" Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.119552 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.125408153 podStartE2EDuration="1m17.119535222s" podCreationTimestamp="2025-10-02 11:12:35 +0000 UTC" firstStartedPulling="2025-10-02 11:12:36.958781484 +0000 UTC m=+1239.013007934" lastFinishedPulling="2025-10-02 11:13:05.952908563 +0000 UTC m=+1268.007135003" observedRunningTime="2025-10-02 11:13:52.115785292 +0000 UTC m=+1314.170011752" watchObservedRunningTime="2025-10-02 11:13:52.119535222 +0000 UTC m=+1314.173761672" Oct 02 11:13:52 crc kubenswrapper[4751]: I1002 11:13:52.122079 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 11:13:52 crc kubenswrapper[4751]: E1002 11:13:52.463456 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Oct 02 11:13:52 crc kubenswrapper[4751]: E1002 11:13:52.463829 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s868k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(6b3aa529-c99f-464f-85a0-70711c59f08f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:13:52 crc kubenswrapper[4751]: E1002 11:13:52.465043 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="6b3aa529-c99f-464f-85a0-70711c59f08f" Oct 02 11:13:53 crc kubenswrapper[4751]: E1002 11:13:53.077609 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="539d6784-c995-47a3-9a80-53223cd9da09" Oct 02 11:13:53 crc kubenswrapper[4751]: E1002 11:13:53.077639 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="6b3aa529-c99f-464f-85a0-70711c59f08f" Oct 02 11:13:54 crc kubenswrapper[4751]: I1002 11:13:54.085078 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerStarted","Data":"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5"} Oct 02 11:13:54 crc kubenswrapper[4751]: I1002 11:13:54.085758 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:54 crc kubenswrapper[4751]: E1002 11:13:54.086471 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="539d6784-c995-47a3-9a80-53223cd9da09" Oct 02 11:13:54 crc kubenswrapper[4751]: I1002 11:13:54.110085 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" podStartSLOduration=23.110064032 podStartE2EDuration="23.110064032s" podCreationTimestamp="2025-10-02 11:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:13:54.106577579 +0000 UTC m=+1316.160804059" watchObservedRunningTime="2025-10-02 11:13:54.110064032 +0000 UTC m=+1316.164290502" Oct 02 11:13:54 crc kubenswrapper[4751]: I1002 11:13:54.661281 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 11:13:54 crc kubenswrapper[4751]: E1002 11:13:54.663752 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="6b3aa529-c99f-464f-85a0-70711c59f08f" Oct 02 11:13:54 crc kubenswrapper[4751]: I1002 11:13:54.972419 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tqqkd" podUID="868943fd-a908-4e94-bef5-737ba8ebc1da" containerName="ovn-controller" probeResult="failure" output=< Oct 02 11:13:54 crc kubenswrapper[4751]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 11:13:54 crc kubenswrapper[4751]: > Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.042804 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.059778 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t54r5" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.102674 4751 generic.go:334] "Generic (PLEG): container finished" podID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerID="3432c82b9e0622ec5f5ece1b81a44963f17f80a50ec2661d0b17a2e1ba4c5a3a" exitCode=0 Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.102787 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerDied","Data":"3432c82b9e0622ec5f5ece1b81a44963f17f80a50ec2661d0b17a2e1ba4c5a3a"} Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.209440 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-jsrq2"] Oct 02 11:13:55 crc kubenswrapper[4751]: E1002 11:13:55.209780 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="init" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.209797 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="init" Oct 02 11:13:55 crc kubenswrapper[4751]: E1002 11:13:55.209849 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.209858 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.210005 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="53608afd-2a0e-4f7f-ae34-ab1c2848b530" containerName="dnsmasq-dns" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.210547 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.212833 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.227030 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jsrq2"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332031 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85a9e69-5573-4806-ab4b-0b10c57d9c91-config\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332219 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn7k9\" (UniqueName: \"kubernetes.io/projected/a85a9e69-5573-4806-ab4b-0b10c57d9c91-kube-api-access-vn7k9\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332257 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-combined-ca-bundle\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332279 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovs-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332625 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovn-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.332760 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434520 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn7k9\" (UniqueName: \"kubernetes.io/projected/a85a9e69-5573-4806-ab4b-0b10c57d9c91-kube-api-access-vn7k9\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434583 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovs-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434606 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-combined-ca-bundle\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434701 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovn-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434748 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434806 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85a9e69-5573-4806-ab4b-0b10c57d9c91-config\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.434949 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovs-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.435002 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a85a9e69-5573-4806-ab4b-0b10c57d9c91-ovn-rundir\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.435629 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a85a9e69-5573-4806-ab4b-0b10c57d9c91-config\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.453396 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.453490 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85a9e69-5573-4806-ab4b-0b10c57d9c91-combined-ca-bundle\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.468980 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn7k9\" (UniqueName: \"kubernetes.io/projected/a85a9e69-5573-4806-ab4b-0b10c57d9c91-kube-api-access-vn7k9\") pod \"ovn-controller-metrics-jsrq2\" (UID: \"a85a9e69-5573-4806-ab4b-0b10c57d9c91\") " pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.536275 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jsrq2" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.543263 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.572411 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-w6fl7"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.574207 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.580634 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.589303 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-w6fl7"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.661914 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 11:13:55 crc kubenswrapper[4751]: E1002 11:13:55.665104 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="6b3aa529-c99f-464f-85a0-70711c59f08f" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.718886 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.721348 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-w6fl7"] Oct 02 11:13:55 crc kubenswrapper[4751]: E1002 11:13:55.721898 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-4cgst ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" podUID="366ed9c6-c289-4dfe-982f-a213ffb49087" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.740247 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.740329 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.740382 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cgst\" (UniqueName: \"kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.740550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.763947 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.765916 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.772486 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.774203 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.841811 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.841897 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cgst\" (UniqueName: \"kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.841982 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.842045 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.842624 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.842996 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.843120 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.860174 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cgst\" (UniqueName: \"kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst\") pod \"dnsmasq-dns-57d65f699f-w6fl7\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.943339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.943756 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.943791 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpzsw\" (UniqueName: \"kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.943833 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:55 crc kubenswrapper[4751]: I1002 11:13:55.943914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.045100 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.045172 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.045228 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpzsw\" (UniqueName: \"kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.045273 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.045350 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.047048 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.047218 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.047252 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.049519 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.064543 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpzsw\" (UniqueName: \"kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw\") pod \"dnsmasq-dns-b8fbc5445-khsms\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.084415 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.128182 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerStarted","Data":"91fe3fd870d8dddb3232de4e25185991bcc93aa0b82d75d6afe4d2c5efceafb2"} Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.128973 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.129660 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="dnsmasq-dns" containerID="cri-o://946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5" gracePeriod=10 Oct 02 11:13:56 crc kubenswrapper[4751]: E1002 11:13:56.130439 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="6b3aa529-c99f-464f-85a0-70711c59f08f" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.139804 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.169171 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371955.68563 podStartE2EDuration="1m21.169144285s" podCreationTimestamp="2025-10-02 11:12:35 +0000 UTC" firstStartedPulling="2025-10-02 11:12:37.23133142 +0000 UTC m=+1239.285557870" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:13:56.168664943 +0000 UTC m=+1318.222891393" watchObservedRunningTime="2025-10-02 11:13:56.169144285 +0000 UTC m=+1318.223370735" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.190473 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.247764 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc\") pod \"366ed9c6-c289-4dfe-982f-a213ffb49087\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.248225 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb\") pod \"366ed9c6-c289-4dfe-982f-a213ffb49087\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.248287 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cgst\" (UniqueName: \"kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst\") pod \"366ed9c6-c289-4dfe-982f-a213ffb49087\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.248395 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config\") pod \"366ed9c6-c289-4dfe-982f-a213ffb49087\" (UID: \"366ed9c6-c289-4dfe-982f-a213ffb49087\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.248825 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "366ed9c6-c289-4dfe-982f-a213ffb49087" (UID: "366ed9c6-c289-4dfe-982f-a213ffb49087"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.249233 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config" (OuterVolumeSpecName: "config") pod "366ed9c6-c289-4dfe-982f-a213ffb49087" (UID: "366ed9c6-c289-4dfe-982f-a213ffb49087"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.248572 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "366ed9c6-c289-4dfe-982f-a213ffb49087" (UID: "366ed9c6-c289-4dfe-982f-a213ffb49087"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.249950 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.249977 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.249988 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/366ed9c6-c289-4dfe-982f-a213ffb49087-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.253565 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst" (OuterVolumeSpecName: "kube-api-access-4cgst") pod "366ed9c6-c289-4dfe-982f-a213ffb49087" (UID: "366ed9c6-c289-4dfe-982f-a213ffb49087"). InnerVolumeSpecName "kube-api-access-4cgst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.280494 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jsrq2"] Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.351162 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cgst\" (UniqueName: \"kubernetes.io/projected/366ed9c6-c289-4dfe-982f-a213ffb49087-kube-api-access-4cgst\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.604170 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:13:56 crc kubenswrapper[4751]: W1002 11:13:56.618402 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55adacb7_50c1_41b0_9c40_171fdae99382.slice/crio-c39f89468c39b91c965a9d6b916a52d8a67f1d2844d8c6ca21dd03380ee2bf8c WatchSource:0}: Error finding container c39f89468c39b91c965a9d6b916a52d8a67f1d2844d8c6ca21dd03380ee2bf8c: Status 404 returned error can't find the container with id c39f89468c39b91c965a9d6b916a52d8a67f1d2844d8c6ca21dd03380ee2bf8c Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.714289 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.741113 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.863242 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs78s\" (UniqueName: \"kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s\") pod \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.863302 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc\") pod \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.863345 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config\") pod \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\" (UID: \"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe\") " Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.868179 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s" (OuterVolumeSpecName: "kube-api-access-zs78s") pod "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" (UID: "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe"). InnerVolumeSpecName "kube-api-access-zs78s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.901294 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" (UID: "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.905223 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config" (OuterVolumeSpecName: "config") pod "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" (UID: "d3814782-f0fa-40c9-a5ab-2b6d4fda91fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.966030 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs78s\" (UniqueName: \"kubernetes.io/projected/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-kube-api-access-zs78s\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.966099 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:56 crc kubenswrapper[4751]: I1002 11:13:56.966109 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.137267 4751 generic.go:334] "Generic (PLEG): container finished" podID="6bd070ec-aa1f-4991-a239-df38cfff5fcf" containerID="178ea8c0ed18701c2772622ccb984118819d36cec4a35073620b9e266ce05e63" exitCode=0 Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.137475 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6bd070ec-aa1f-4991-a239-df38cfff5fcf","Type":"ContainerDied","Data":"178ea8c0ed18701c2772622ccb984118819d36cec4a35073620b9e266ce05e63"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.142584 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jsrq2" event={"ID":"a85a9e69-5573-4806-ab4b-0b10c57d9c91","Type":"ContainerStarted","Data":"9d0a991777ed06ba977cd192849b7a0e18bb76199eb5ee7f250e99abe81b6649"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.142667 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jsrq2" event={"ID":"a85a9e69-5573-4806-ab4b-0b10c57d9c91","Type":"ContainerStarted","Data":"c3e761b81a6eb85da14605df6c30c8a1cffbfa464297ccc1d847d8c8f28c1b00"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.145935 4751 generic.go:334] "Generic (PLEG): container finished" podID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerID="946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5" exitCode=0 Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.146036 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerDied","Data":"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.146080 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" event={"ID":"d3814782-f0fa-40c9-a5ab-2b6d4fda91fe","Type":"ContainerDied","Data":"868fd492a6de2d2549a923feecd1b493860917f8ec605876c8ff4ac948f2868f"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.146120 4751 scope.go:117] "RemoveContainer" containerID="946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.146304 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fznz8" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.148778 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bkkpx" event={"ID":"30a668e4-d395-4b78-a014-279fd35ed304","Type":"ContainerStarted","Data":"548f74cbf9962009af671fda979df069f739c101a6731077ce097abac834f6fa"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.150413 4751 generic.go:334] "Generic (PLEG): container finished" podID="55adacb7-50c1-41b0-9c40-171fdae99382" containerID="0e4d360ae32c0344eb667c53bcd62dc29184ad6259acc91d15468afa34dc6492" exitCode=0 Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.150442 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" event={"ID":"55adacb7-50c1-41b0-9c40-171fdae99382","Type":"ContainerDied","Data":"0e4d360ae32c0344eb667c53bcd62dc29184ad6259acc91d15468afa34dc6492"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.150463 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" event={"ID":"55adacb7-50c1-41b0-9c40-171fdae99382","Type":"ContainerStarted","Data":"c39f89468c39b91c965a9d6b916a52d8a67f1d2844d8c6ca21dd03380ee2bf8c"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.152590 4751 generic.go:334] "Generic (PLEG): container finished" podID="fa2f2972-7b83-44e7-8ba5-37fb04475f5c" containerID="30ff247f02bb4f624c9c0248804548ee427f55915812bac78daf34811a60d8d6" exitCode=0 Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.152936 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fa2f2972-7b83-44e7-8ba5-37fb04475f5c","Type":"ContainerDied","Data":"30ff247f02bb4f624c9c0248804548ee427f55915812bac78daf34811a60d8d6"} Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.153625 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-w6fl7" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.183226 4751 scope.go:117] "RemoveContainer" containerID="f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.194602 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bkkpx" podStartSLOduration=12.780671406 podStartE2EDuration="21.194584096s" podCreationTimestamp="2025-10-02 11:13:36 +0000 UTC" firstStartedPulling="2025-10-02 11:13:47.515308022 +0000 UTC m=+1309.569534472" lastFinishedPulling="2025-10-02 11:13:55.929220712 +0000 UTC m=+1317.983447162" observedRunningTime="2025-10-02 11:13:57.191259088 +0000 UTC m=+1319.245485598" watchObservedRunningTime="2025-10-02 11:13:57.194584096 +0000 UTC m=+1319.248810546" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.231602 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-jsrq2" podStartSLOduration=1.7292652579999999 podStartE2EDuration="2.231582596s" podCreationTimestamp="2025-10-02 11:13:55 +0000 UTC" firstStartedPulling="2025-10-02 11:13:56.290243543 +0000 UTC m=+1318.344469993" lastFinishedPulling="2025-10-02 11:13:56.792560881 +0000 UTC m=+1318.846787331" observedRunningTime="2025-10-02 11:13:57.219473572 +0000 UTC m=+1319.273700022" watchObservedRunningTime="2025-10-02 11:13:57.231582596 +0000 UTC m=+1319.285809046" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.238380 4751 scope.go:117] "RemoveContainer" containerID="946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5" Oct 02 11:13:57 crc kubenswrapper[4751]: E1002 11:13:57.240725 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5\": container with ID starting with 946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5 not found: ID does not exist" containerID="946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.240770 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5"} err="failed to get container status \"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5\": rpc error: code = NotFound desc = could not find container \"946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5\": container with ID starting with 946514c94be93b7a3a3038eab42373d9909337267cbd83a05ab48e5e771bffe5 not found: ID does not exist" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.240799 4751 scope.go:117] "RemoveContainer" containerID="f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd" Oct 02 11:13:57 crc kubenswrapper[4751]: E1002 11:13:57.246081 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd\": container with ID starting with f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd not found: ID does not exist" containerID="f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.246151 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd"} err="failed to get container status \"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd\": rpc error: code = NotFound desc = could not find container \"f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd\": container with ID starting with f1a8423e20c3f6f4527da0127a77ed83efa5738bac591d85847df207cbe489dd not found: ID does not exist" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.334417 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-w6fl7"] Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.342160 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-w6fl7"] Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.349029 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.354245 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fznz8"] Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.562778 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366ed9c6-c289-4dfe-982f-a213ffb49087" path="/var/lib/kubelet/pods/366ed9c6-c289-4dfe-982f-a213ffb49087/volumes" Oct 02 11:13:57 crc kubenswrapper[4751]: I1002 11:13:57.563270 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" path="/var/lib/kubelet/pods/d3814782-f0fa-40c9-a5ab-2b6d4fda91fe/volumes" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.161966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" event={"ID":"55adacb7-50c1-41b0-9c40-171fdae99382","Type":"ContainerStarted","Data":"d985882ebbf19e74598c046f77e673a78cbeb084ac02d84f0e6b16e9c272f4c9"} Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.162393 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.163998 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"fa2f2972-7b83-44e7-8ba5-37fb04475f5c","Type":"ContainerStarted","Data":"b6a2b29f6116a56d06a173c87fb278066142366b4a738695eba35a826d2cc643"} Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.166259 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6b3aa529-c99f-464f-85a0-70711c59f08f","Type":"ContainerStarted","Data":"9e8cb3006ed6166db036c77d872c5ee60c5f77c80abc2cea18f1710f323305ab"} Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.168577 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6bd070ec-aa1f-4991-a239-df38cfff5fcf","Type":"ContainerStarted","Data":"901459f6a89f0a07da8bf2ec8f0b5b20177767cab5dabaecd1fbd26381941ba2"} Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.194320 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" podStartSLOduration=3.194293891 podStartE2EDuration="3.194293891s" podCreationTimestamp="2025-10-02 11:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:13:58.184936731 +0000 UTC m=+1320.239163191" watchObservedRunningTime="2025-10-02 11:13:58.194293891 +0000 UTC m=+1320.248520361" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.231910 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=68.118879002 podStartE2EDuration="1m21.231872525s" podCreationTimestamp="2025-10-02 11:12:37 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.41551976 +0000 UTC m=+1268.469746210" lastFinishedPulling="2025-10-02 11:13:19.528513283 +0000 UTC m=+1281.582739733" observedRunningTime="2025-10-02 11:13:58.207628987 +0000 UTC m=+1320.261855467" watchObservedRunningTime="2025-10-02 11:13:58.231872525 +0000 UTC m=+1320.286098975" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.239872 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=67.945050684 podStartE2EDuration="1m21.239836468s" podCreationTimestamp="2025-10-02 11:12:37 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.233619377 +0000 UTC m=+1268.287845827" lastFinishedPulling="2025-10-02 11:13:19.528405121 +0000 UTC m=+1281.582631611" observedRunningTime="2025-10-02 11:13:58.23614928 +0000 UTC m=+1320.290375740" watchObservedRunningTime="2025-10-02 11:13:58.239836468 +0000 UTC m=+1320.294062918" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.273444 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=59.391992905 podStartE2EDuration="1m11.273421746s" podCreationTimestamp="2025-10-02 11:12:47 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.494157942 +0000 UTC m=+1268.548384392" lastFinishedPulling="2025-10-02 11:13:18.375586753 +0000 UTC m=+1280.429813233" observedRunningTime="2025-10-02 11:13:58.263515621 +0000 UTC m=+1320.317742081" watchObservedRunningTime="2025-10-02 11:13:58.273421746 +0000 UTC m=+1320.327648216" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.446592 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tqqkd-config-8fj44"] Oct 02 11:13:58 crc kubenswrapper[4751]: E1002 11:13:58.446980 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="dnsmasq-dns" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.447004 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="dnsmasq-dns" Oct 02 11:13:58 crc kubenswrapper[4751]: E1002 11:13:58.447022 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="init" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.447028 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="init" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.447205 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3814782-f0fa-40c9-a5ab-2b6d4fda91fe" containerName="dnsmasq-dns" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.447709 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.449460 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.449487 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.449554 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.462403 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tqqkd-config-8fj44"] Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.596890 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.597339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.597472 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.597546 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.597572 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.597641 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8sxr\" (UniqueName: \"kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698035 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698109 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698139 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698154 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698210 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8sxr\" (UniqueName: \"kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698239 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698509 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.698571 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.699224 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.699277 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.701232 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.733496 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8sxr\" (UniqueName: \"kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr\") pod \"ovn-controller-tqqkd-config-8fj44\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:58 crc kubenswrapper[4751]: I1002 11:13:58.767221 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:13:59 crc kubenswrapper[4751]: I1002 11:13:59.214466 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 11:13:59 crc kubenswrapper[4751]: I1002 11:13:59.214822 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 11:13:59 crc kubenswrapper[4751]: I1002 11:13:59.227553 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tqqkd-config-8fj44"] Oct 02 11:13:59 crc kubenswrapper[4751]: I1002 11:13:59.970750 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tqqkd" Oct 02 11:14:00 crc kubenswrapper[4751]: I1002 11:14:00.190334 4751 generic.go:334] "Generic (PLEG): container finished" podID="94f6773e-9dc7-4116-97ba-f364113ca761" containerID="c4b1ba3fc0d08e4a2c9125ebe704a9e6704ff9971f808dc396e3a11ea62832b8" exitCode=0 Oct 02 11:14:00 crc kubenswrapper[4751]: I1002 11:14:00.190394 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tqqkd-config-8fj44" event={"ID":"94f6773e-9dc7-4116-97ba-f364113ca761","Type":"ContainerDied","Data":"c4b1ba3fc0d08e4a2c9125ebe704a9e6704ff9971f808dc396e3a11ea62832b8"} Oct 02 11:14:00 crc kubenswrapper[4751]: I1002 11:14:00.190423 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tqqkd-config-8fj44" event={"ID":"94f6773e-9dc7-4116-97ba-f364113ca761","Type":"ContainerStarted","Data":"94e1b650da2574b59bf9b0ef1ad2fad2c9acbe42d6f3a658d4f40b9040ed2e12"} Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.506744 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.507022 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.547466 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.645721 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.645797 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.645856 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.645888 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.645978 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646036 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646046 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646083 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run" (OuterVolumeSpecName: "var-run") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646196 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8sxr\" (UniqueName: \"kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr\") pod \"94f6773e-9dc7-4116-97ba-f364113ca761\" (UID: \"94f6773e-9dc7-4116-97ba-f364113ca761\") " Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646721 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646755 4751 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646817 4751 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646833 4751 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/94f6773e-9dc7-4116-97ba-f364113ca761-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.646905 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts" (OuterVolumeSpecName: "scripts") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.651357 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr" (OuterVolumeSpecName: "kube-api-access-n8sxr") pod "94f6773e-9dc7-4116-97ba-f364113ca761" (UID: "94f6773e-9dc7-4116-97ba-f364113ca761"). InnerVolumeSpecName "kube-api-access-n8sxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.748058 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8sxr\" (UniqueName: \"kubernetes.io/projected/94f6773e-9dc7-4116-97ba-f364113ca761-kube-api-access-n8sxr\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.748091 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:01 crc kubenswrapper[4751]: I1002 11:14:01.748116 4751 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/94f6773e-9dc7-4116-97ba-f364113ca761-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.209655 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tqqkd-config-8fj44" event={"ID":"94f6773e-9dc7-4116-97ba-f364113ca761","Type":"ContainerDied","Data":"94e1b650da2574b59bf9b0ef1ad2fad2c9acbe42d6f3a658d4f40b9040ed2e12"} Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.209959 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94e1b650da2574b59bf9b0ef1ad2fad2c9acbe42d6f3a658d4f40b9040ed2e12" Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.209719 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tqqkd-config-8fj44" Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.538740 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.606742 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.677674 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tqqkd-config-8fj44"] Oct 02 11:14:02 crc kubenswrapper[4751]: I1002 11:14:02.689886 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tqqkd-config-8fj44"] Oct 02 11:14:03 crc kubenswrapper[4751]: I1002 11:14:03.562728 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94f6773e-9dc7-4116-97ba-f364113ca761" path="/var/lib/kubelet/pods/94f6773e-9dc7-4116-97ba-f364113ca761/volumes" Oct 02 11:14:04 crc kubenswrapper[4751]: I1002 11:14:04.698521 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:14:04 crc kubenswrapper[4751]: I1002 11:14:04.707525 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0071ad0-2497-45b8-bddc-9164ee537d18-etc-swift\") pod \"swift-storage-0\" (UID: \"b0071ad0-2497-45b8-bddc-9164ee537d18\") " pod="openstack/swift-storage-0" Oct 02 11:14:04 crc kubenswrapper[4751]: I1002 11:14:04.895727 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 11:14:05 crc kubenswrapper[4751]: I1002 11:14:05.235628 4751 generic.go:334] "Generic (PLEG): container finished" podID="30a668e4-d395-4b78-a014-279fd35ed304" containerID="548f74cbf9962009af671fda979df069f739c101a6731077ce097abac834f6fa" exitCode=0 Oct 02 11:14:05 crc kubenswrapper[4751]: I1002 11:14:05.235669 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bkkpx" event={"ID":"30a668e4-d395-4b78-a014-279fd35ed304","Type":"ContainerDied","Data":"548f74cbf9962009af671fda979df069f739c101a6731077ce097abac834f6fa"} Oct 02 11:14:05 crc kubenswrapper[4751]: I1002 11:14:05.486369 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 11:14:05 crc kubenswrapper[4751]: I1002 11:14:05.877441 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 11:14:05 crc kubenswrapper[4751]: I1002 11:14:05.937815 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.086804 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.132509 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.132775 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="dnsmasq-dns" containerID="cri-o://16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0" gracePeriod=10 Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.255761 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"4b68f566631f40e94d7fd39c945cb9a4d384a191f1fc1b1caebe82904570a464"} Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.469253 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.695307 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.751932 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.840953 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841018 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841126 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfgq2\" (UniqueName: \"kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841192 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841539 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841593 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.841698 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices\") pod \"30a668e4-d395-4b78-a014-279fd35ed304\" (UID: \"30a668e4-d395-4b78-a014-279fd35ed304\") " Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.843248 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.843621 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.865494 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2" (OuterVolumeSpecName: "kube-api-access-hfgq2") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "kube-api-access-hfgq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.867985 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts" (OuterVolumeSpecName: "scripts") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.868097 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.874424 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.876250 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "30a668e4-d395-4b78-a014-279fd35ed304" (UID: "30a668e4-d395-4b78-a014-279fd35ed304"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.937087 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-ffs5s"] Oct 02 11:14:06 crc kubenswrapper[4751]: E1002 11:14:06.937853 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a668e4-d395-4b78-a014-279fd35ed304" containerName="swift-ring-rebalance" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.937878 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a668e4-d395-4b78-a014-279fd35ed304" containerName="swift-ring-rebalance" Oct 02 11:14:06 crc kubenswrapper[4751]: E1002 11:14:06.937903 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f6773e-9dc7-4116-97ba-f364113ca761" containerName="ovn-config" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.937911 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f6773e-9dc7-4116-97ba-f364113ca761" containerName="ovn-config" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.938608 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f6773e-9dc7-4116-97ba-f364113ca761" containerName="ovn-config" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.938679 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a668e4-d395-4b78-a014-279fd35ed304" containerName="swift-ring-rebalance" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.939805 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944453 4751 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944497 4751 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944513 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30a668e4-d395-4b78-a014-279fd35ed304-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944528 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfgq2\" (UniqueName: \"kubernetes.io/projected/30a668e4-d395-4b78-a014-279fd35ed304-kube-api-access-hfgq2\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944547 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944560 4751 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30a668e4-d395-4b78-a014-279fd35ed304-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.944571 4751 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30a668e4-d395-4b78-a014-279fd35ed304-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:06 crc kubenswrapper[4751]: I1002 11:14:06.948573 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ffs5s"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.023553 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qzv47"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.024606 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.031660 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qzv47"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.049113 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glvfz\" (UniqueName: \"kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz\") pod \"barbican-db-create-ffs5s\" (UID: \"6514a959-a0fd-4226-8676-00c411e7e1fb\") " pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.151515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glvfz\" (UniqueName: \"kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz\") pod \"barbican-db-create-ffs5s\" (UID: \"6514a959-a0fd-4226-8676-00c411e7e1fb\") " pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.151621 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbktj\" (UniqueName: \"kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj\") pod \"cinder-db-create-qzv47\" (UID: \"39a3d249-497a-4ea8-9752-6f21d35f392d\") " pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.172529 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glvfz\" (UniqueName: \"kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz\") pod \"barbican-db-create-ffs5s\" (UID: \"6514a959-a0fd-4226-8676-00c411e7e1fb\") " pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.221430 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.256033 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbktj\" (UniqueName: \"kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj\") pod \"cinder-db-create-qzv47\" (UID: \"39a3d249-497a-4ea8-9752-6f21d35f392d\") " pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.267793 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.286393 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbktj\" (UniqueName: \"kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj\") pod \"cinder-db-create-qzv47\" (UID: \"39a3d249-497a-4ea8-9752-6f21d35f392d\") " pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.294873 4751 generic.go:334] "Generic (PLEG): container finished" podID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerID="16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0" exitCode=0 Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.294944 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" event={"ID":"9991d5e7-fcb2-4535-b9c8-d327ef696edc","Type":"ContainerDied","Data":"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0"} Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.294972 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" event={"ID":"9991d5e7-fcb2-4535-b9c8-d327ef696edc","Type":"ContainerDied","Data":"7fd1b4cf94bcb4b84f792e856647c845b948f68f064e7ed0f966eab35c0615df"} Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.294988 4751 scope.go:117] "RemoveContainer" containerID="16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.295103 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-2tdxk" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.308759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bkkpx" event={"ID":"30a668e4-d395-4b78-a014-279fd35ed304","Type":"ContainerDied","Data":"9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde"} Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.308810 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ec9094956b49132acf345bd481db0d2ade93ac5d685799389a89478312bcbde" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.308912 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bkkpx" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.345581 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.366029 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc\") pod \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.366096 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config\") pod \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.366168 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kglz\" (UniqueName: \"kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz\") pod \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\" (UID: \"9991d5e7-fcb2-4535-b9c8-d327ef696edc\") " Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.377027 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz" (OuterVolumeSpecName: "kube-api-access-5kglz") pod "9991d5e7-fcb2-4535-b9c8-d327ef696edc" (UID: "9991d5e7-fcb2-4535-b9c8-d327ef696edc"). InnerVolumeSpecName "kube-api-access-5kglz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.387958 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-fg7cq"] Oct 02 11:14:07 crc kubenswrapper[4751]: E1002 11:14:07.424717 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="init" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.424753 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="init" Oct 02 11:14:07 crc kubenswrapper[4751]: E1002 11:14:07.424774 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="dnsmasq-dns" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.424781 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="dnsmasq-dns" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.425033 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" containerName="dnsmasq-dns" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.425673 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.442003 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fg7cq"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.455032 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config" (OuterVolumeSpecName: "config") pod "9991d5e7-fcb2-4535-b9c8-d327ef696edc" (UID: "9991d5e7-fcb2-4535-b9c8-d327ef696edc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.467644 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.467669 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kglz\" (UniqueName: \"kubernetes.io/projected/9991d5e7-fcb2-4535-b9c8-d327ef696edc-kube-api-access-5kglz\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.494923 4751 scope.go:117] "RemoveContainer" containerID="8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.498750 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9991d5e7-fcb2-4535-b9c8-d327ef696edc" (UID: "9991d5e7-fcb2-4535-b9c8-d327ef696edc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.533825 4751 scope.go:117] "RemoveContainer" containerID="16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0" Oct 02 11:14:07 crc kubenswrapper[4751]: E1002 11:14:07.535788 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0\": container with ID starting with 16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0 not found: ID does not exist" containerID="16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.535827 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0"} err="failed to get container status \"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0\": rpc error: code = NotFound desc = could not find container \"16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0\": container with ID starting with 16f97679fe6d446f0bb167265654b7f6b470a4a91ee49f75b4ea8f2e46eec5f0 not found: ID does not exist" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.535856 4751 scope.go:117] "RemoveContainer" containerID="8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d" Oct 02 11:14:07 crc kubenswrapper[4751]: E1002 11:14:07.552459 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d\": container with ID starting with 8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d not found: ID does not exist" containerID="8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.552759 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d"} err="failed to get container status \"8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d\": rpc error: code = NotFound desc = could not find container \"8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d\": container with ID starting with 8af00df56ccb2d768db9608c2b2309c4ef0e04b9930a06b2f8e8232319bcf12d not found: ID does not exist" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.579895 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlp7\" (UniqueName: \"kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7\") pod \"neutron-db-create-fg7cq\" (UID: \"bc55b82d-1e7c-4036-b372-e881a31dcbe5\") " pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.580034 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9991d5e7-fcb2-4535-b9c8-d327ef696edc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.678538 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.682823 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlp7\" (UniqueName: \"kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7\") pod \"neutron-db-create-fg7cq\" (UID: \"bc55b82d-1e7c-4036-b372-e881a31dcbe5\") " pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.694192 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-2tdxk"] Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.707367 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlp7\" (UniqueName: \"kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7\") pod \"neutron-db-create-fg7cq\" (UID: \"bc55b82d-1e7c-4036-b372-e881a31dcbe5\") " pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.744457 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:07 crc kubenswrapper[4751]: I1002 11:14:07.971832 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ffs5s"] Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.019071 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-fg7cq"] Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.073335 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qzv47"] Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.319114 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fg7cq" event={"ID":"bc55b82d-1e7c-4036-b372-e881a31dcbe5","Type":"ContainerStarted","Data":"f8f7debf23638d626b3696018573cf1f3e92dcd55cae289f826fd0af2751a55a"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.323444 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"8aaf1dbf89321aa9b0ada65ec47e887ec729c361ec66c12abf509cdc9625565e"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.323482 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"e4b449098af8ee0b4644f147777e641e4103e6e880bac67b380717abf05a56ab"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.324797 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ffs5s" event={"ID":"6514a959-a0fd-4226-8676-00c411e7e1fb","Type":"ContainerStarted","Data":"f69d701189e38061c3c231ce6eaa3b89c7d263ca7931449417c52e92a5125ba0"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.328057 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"539d6784-c995-47a3-9a80-53223cd9da09","Type":"ContainerStarted","Data":"633d97cba19de97d24592cc39c61cb996a22427af8c645a1719e730b77e6ca70"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.329586 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qzv47" event={"ID":"39a3d249-497a-4ea8-9752-6f21d35f392d","Type":"ContainerStarted","Data":"7e51cd0dee3fe2417c57ab65254a04610ba0ceafdf7818f41f46aae75e76eb76"} Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.634053 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=71.60705164 podStartE2EDuration="1m24.634032425s" podCreationTimestamp="2025-10-02 11:12:44 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.686562835 +0000 UTC m=+1268.740789285" lastFinishedPulling="2025-10-02 11:13:19.71354362 +0000 UTC m=+1281.767770070" observedRunningTime="2025-10-02 11:14:08.360444891 +0000 UTC m=+1330.414671351" watchObservedRunningTime="2025-10-02 11:14:08.634032425 +0000 UTC m=+1330.688258875" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.649470 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.651093 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.653859 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.654090 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.654297 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-cdfgb" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.654454 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.657441 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803422 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv6zt\" (UniqueName: \"kubernetes.io/projected/53ba58f6-e55e-4bd6-9365-04fe987aec95-kube-api-access-cv6zt\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803557 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803599 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803638 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803659 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803692 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-config\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.803914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-scripts\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.905159 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.905239 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.905265 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.905280 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.905305 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-config\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.906101 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.906273 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-scripts\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.906319 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv6zt\" (UniqueName: \"kubernetes.io/projected/53ba58f6-e55e-4bd6-9365-04fe987aec95-kube-api-access-cv6zt\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.907195 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-config\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.907617 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ba58f6-e55e-4bd6-9365-04fe987aec95-scripts\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.912609 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.925718 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv6zt\" (UniqueName: \"kubernetes.io/projected/53ba58f6-e55e-4bd6-9365-04fe987aec95-kube-api-access-cv6zt\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.925992 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.927503 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ba58f6-e55e-4bd6-9365-04fe987aec95-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"53ba58f6-e55e-4bd6-9365-04fe987aec95\") " pod="openstack/ovn-northd-0" Oct 02 11:14:08 crc kubenswrapper[4751]: I1002 11:14:08.974604 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.332311 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-l56lv"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.334347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.337339 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qzv47" event={"ID":"39a3d249-497a-4ea8-9752-6f21d35f392d","Type":"ContainerStarted","Data":"0cc9c5351195fb28145a8f3243beba900678730dd228b5fecb6e920247eeda0e"} Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.339075 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fg7cq" event={"ID":"bc55b82d-1e7c-4036-b372-e881a31dcbe5","Type":"ContainerStarted","Data":"68db4c24bc1909258a77a59fbebb97eb8008cd8e86f4a31e31bbc33b18031ec9"} Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.342505 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"fc8f32853d29e18f3bb1a25d1f6571b3973f42dc58abfd9db52b733245de18e3"} Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.342550 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"bfd158c4281f4f442c0c7ae6db6300e0c9c683029eae15e57f9f31ee8f7c2b7a"} Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.345013 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ffs5s" event={"ID":"6514a959-a0fd-4226-8676-00c411e7e1fb","Type":"ContainerStarted","Data":"720bf7cc015540cca238cc11156072acaec8b5c7cf306a86feeb69781eb45d13"} Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.352184 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l56lv"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.390526 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qzv47" podStartSLOduration=3.3905055969999998 podStartE2EDuration="3.390505597s" podCreationTimestamp="2025-10-02 11:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:09.390412765 +0000 UTC m=+1331.444639215" watchObservedRunningTime="2025-10-02 11:14:09.390505597 +0000 UTC m=+1331.444732057" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.414155 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-fg7cq" podStartSLOduration=2.414137319 podStartE2EDuration="2.414137319s" podCreationTimestamp="2025-10-02 11:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:09.410589204 +0000 UTC m=+1331.464815654" watchObservedRunningTime="2025-10-02 11:14:09.414137319 +0000 UTC m=+1331.468363779" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.416270 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2svtz\" (UniqueName: \"kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz\") pod \"keystone-db-create-l56lv\" (UID: \"d45165ce-0577-4093-b5b2-268035e7540f\") " pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.429695 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-ffs5s" podStartSLOduration=3.429672704 podStartE2EDuration="3.429672704s" podCreationTimestamp="2025-10-02 11:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:09.425548654 +0000 UTC m=+1331.479775124" watchObservedRunningTime="2025-10-02 11:14:09.429672704 +0000 UTC m=+1331.483899164" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.466001 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.518196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2svtz\" (UniqueName: \"kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz\") pod \"keystone-db-create-l56lv\" (UID: \"d45165ce-0577-4093-b5b2-268035e7540f\") " pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.537017 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2svtz\" (UniqueName: \"kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz\") pod \"keystone-db-create-l56lv\" (UID: \"d45165ce-0577-4093-b5b2-268035e7540f\") " pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.561207 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9991d5e7-fcb2-4535-b9c8-d327ef696edc" path="/var/lib/kubelet/pods/9991d5e7-fcb2-4535-b9c8-d327ef696edc/volumes" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.662630 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.741970 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-pz7qh"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.743012 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.749615 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pz7qh"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.824367 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djtnv\" (UniqueName: \"kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv\") pod \"placement-db-create-pz7qh\" (UID: \"d9944c9c-cead-44e6-b589-788b4ef04cc7\") " pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.896970 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ll6rj"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.898197 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.914093 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ll6rj"] Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.926506 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djtnv\" (UniqueName: \"kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv\") pod \"placement-db-create-pz7qh\" (UID: \"d9944c9c-cead-44e6-b589-788b4ef04cc7\") " pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:09 crc kubenswrapper[4751]: I1002 11:14:09.944716 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djtnv\" (UniqueName: \"kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv\") pod \"placement-db-create-pz7qh\" (UID: \"d9944c9c-cead-44e6-b589-788b4ef04cc7\") " pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.028463 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zhzl\" (UniqueName: \"kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl\") pod \"glance-db-create-ll6rj\" (UID: \"16d45636-f990-47b9-923f-f1464ea8c046\") " pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.082099 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.133194 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zhzl\" (UniqueName: \"kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl\") pod \"glance-db-create-ll6rj\" (UID: \"16d45636-f990-47b9-923f-f1464ea8c046\") " pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.139793 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l56lv"] Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.150353 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zhzl\" (UniqueName: \"kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl\") pod \"glance-db-create-ll6rj\" (UID: \"16d45636-f990-47b9-923f-f1464ea8c046\") " pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:10 crc kubenswrapper[4751]: W1002 11:14:10.153202 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd45165ce_0577_4093_b5b2_268035e7540f.slice/crio-9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65 WatchSource:0}: Error finding container 9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65: Status 404 returned error can't find the container with id 9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65 Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.223215 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:10 crc kubenswrapper[4751]: E1002 11:14:10.284011 4751 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.163:50992->38.129.56.163:40831: read tcp 38.129.56.163:50992->38.129.56.163:40831: read: connection reset by peer Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.360678 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l56lv" event={"ID":"d45165ce-0577-4093-b5b2-268035e7540f","Type":"ContainerStarted","Data":"9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65"} Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.361859 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53ba58f6-e55e-4bd6-9365-04fe987aec95","Type":"ContainerStarted","Data":"f84dce2274705c8a97891ac9c08026cdc543d03b6df4eb5cf3ed3544fdbc5729"} Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.588894 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pz7qh"] Oct 02 11:14:10 crc kubenswrapper[4751]: I1002 11:14:10.732144 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ll6rj"] Oct 02 11:14:10 crc kubenswrapper[4751]: W1002 11:14:10.738202 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16d45636_f990_47b9_923f_f1464ea8c046.slice/crio-f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593 WatchSource:0}: Error finding container f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593: Status 404 returned error can't find the container with id f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593 Oct 02 11:14:11 crc kubenswrapper[4751]: I1002 11:14:11.369766 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pz7qh" event={"ID":"d9944c9c-cead-44e6-b589-788b4ef04cc7","Type":"ContainerStarted","Data":"0eccdb14731e19bb2fe35c3ce26e16c2e7b5f16aa14631b5923c2961c9ff1fd4"} Oct 02 11:14:11 crc kubenswrapper[4751]: I1002 11:14:11.371933 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ll6rj" event={"ID":"16d45636-f990-47b9-923f-f1464ea8c046","Type":"ContainerStarted","Data":"f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593"} Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.405637 4751 generic.go:334] "Generic (PLEG): container finished" podID="39a3d249-497a-4ea8-9752-6f21d35f392d" containerID="0cc9c5351195fb28145a8f3243beba900678730dd228b5fecb6e920247eeda0e" exitCode=0 Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.405747 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qzv47" event={"ID":"39a3d249-497a-4ea8-9752-6f21d35f392d","Type":"ContainerDied","Data":"0cc9c5351195fb28145a8f3243beba900678730dd228b5fecb6e920247eeda0e"} Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.408873 4751 generic.go:334] "Generic (PLEG): container finished" podID="bc55b82d-1e7c-4036-b372-e881a31dcbe5" containerID="68db4c24bc1909258a77a59fbebb97eb8008cd8e86f4a31e31bbc33b18031ec9" exitCode=0 Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.408943 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fg7cq" event={"ID":"bc55b82d-1e7c-4036-b372-e881a31dcbe5","Type":"ContainerDied","Data":"68db4c24bc1909258a77a59fbebb97eb8008cd8e86f4a31e31bbc33b18031ec9"} Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.410732 4751 generic.go:334] "Generic (PLEG): container finished" podID="6514a959-a0fd-4226-8676-00c411e7e1fb" containerID="720bf7cc015540cca238cc11156072acaec8b5c7cf306a86feeb69781eb45d13" exitCode=0 Oct 02 11:14:15 crc kubenswrapper[4751]: I1002 11:14:15.410758 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ffs5s" event={"ID":"6514a959-a0fd-4226-8676-00c411e7e1fb","Type":"ContainerDied","Data":"720bf7cc015540cca238cc11156072acaec8b5c7cf306a86feeb69781eb45d13"} Oct 02 11:14:16 crc kubenswrapper[4751]: I1002 11:14:16.423236 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pz7qh" event={"ID":"d9944c9c-cead-44e6-b589-788b4ef04cc7","Type":"ContainerStarted","Data":"946b76220a37bd82f78c15f93a3a24ad4369dce8671dc3ebf96c66fecd7ea82e"} Oct 02 11:14:17 crc kubenswrapper[4751]: I1002 11:14:17.454218 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-pz7qh" podStartSLOduration=8.454154714 podStartE2EDuration="8.454154714s" podCreationTimestamp="2025-10-02 11:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:17.452715686 +0000 UTC m=+1339.506942136" watchObservedRunningTime="2025-10-02 11:14:17.454154714 +0000 UTC m=+1339.508381204" Oct 02 11:14:19 crc kubenswrapper[4751]: I1002 11:14:19.447389 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ll6rj" event={"ID":"16d45636-f990-47b9-923f-f1464ea8c046","Type":"ContainerStarted","Data":"e96ef46ae2fdba975bd6990dbc6007d8d60433914e44c082acaa0920958de230"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.134136 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.204139 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbktj\" (UniqueName: \"kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj\") pod \"39a3d249-497a-4ea8-9752-6f21d35f392d\" (UID: \"39a3d249-497a-4ea8-9752-6f21d35f392d\") " Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.211470 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj" (OuterVolumeSpecName: "kube-api-access-hbktj") pod "39a3d249-497a-4ea8-9752-6f21d35f392d" (UID: "39a3d249-497a-4ea8-9752-6f21d35f392d"). InnerVolumeSpecName "kube-api-access-hbktj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.305797 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.306665 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbktj\" (UniqueName: \"kubernetes.io/projected/39a3d249-497a-4ea8-9752-6f21d35f392d-kube-api-access-hbktj\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.312297 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.407567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwlp7\" (UniqueName: \"kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7\") pod \"bc55b82d-1e7c-4036-b372-e881a31dcbe5\" (UID: \"bc55b82d-1e7c-4036-b372-e881a31dcbe5\") " Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.407708 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glvfz\" (UniqueName: \"kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz\") pod \"6514a959-a0fd-4226-8676-00c411e7e1fb\" (UID: \"6514a959-a0fd-4226-8676-00c411e7e1fb\") " Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.412501 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7" (OuterVolumeSpecName: "kube-api-access-fwlp7") pod "bc55b82d-1e7c-4036-b372-e881a31dcbe5" (UID: "bc55b82d-1e7c-4036-b372-e881a31dcbe5"). InnerVolumeSpecName "kube-api-access-fwlp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.414558 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz" (OuterVolumeSpecName: "kube-api-access-glvfz") pod "6514a959-a0fd-4226-8676-00c411e7e1fb" (UID: "6514a959-a0fd-4226-8676-00c411e7e1fb"). InnerVolumeSpecName "kube-api-access-glvfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.457793 4751 generic.go:334] "Generic (PLEG): container finished" podID="d9944c9c-cead-44e6-b589-788b4ef04cc7" containerID="946b76220a37bd82f78c15f93a3a24ad4369dce8671dc3ebf96c66fecd7ea82e" exitCode=0 Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.457853 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pz7qh" event={"ID":"d9944c9c-cead-44e6-b589-788b4ef04cc7","Type":"ContainerDied","Data":"946b76220a37bd82f78c15f93a3a24ad4369dce8671dc3ebf96c66fecd7ea82e"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.463209 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ffs5s" event={"ID":"6514a959-a0fd-4226-8676-00c411e7e1fb","Type":"ContainerDied","Data":"f69d701189e38061c3c231ce6eaa3b89c7d263ca7931449417c52e92a5125ba0"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.463298 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f69d701189e38061c3c231ce6eaa3b89c7d263ca7931449417c52e92a5125ba0" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.463220 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ffs5s" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.473924 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qzv47" event={"ID":"39a3d249-497a-4ea8-9752-6f21d35f392d","Type":"ContainerDied","Data":"7e51cd0dee3fe2417c57ab65254a04610ba0ceafdf7818f41f46aae75e76eb76"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.473970 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e51cd0dee3fe2417c57ab65254a04610ba0ceafdf7818f41f46aae75e76eb76" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.475191 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qzv47" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.481457 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-fg7cq" event={"ID":"bc55b82d-1e7c-4036-b372-e881a31dcbe5","Type":"ContainerDied","Data":"f8f7debf23638d626b3696018573cf1f3e92dcd55cae289f826fd0af2751a55a"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.481502 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8f7debf23638d626b3696018573cf1f3e92dcd55cae289f826fd0af2751a55a" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.481569 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-fg7cq" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.487524 4751 generic.go:334] "Generic (PLEG): container finished" podID="16d45636-f990-47b9-923f-f1464ea8c046" containerID="e96ef46ae2fdba975bd6990dbc6007d8d60433914e44c082acaa0920958de230" exitCode=0 Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.487610 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ll6rj" event={"ID":"16d45636-f990-47b9-923f-f1464ea8c046","Type":"ContainerDied","Data":"e96ef46ae2fdba975bd6990dbc6007d8d60433914e44c082acaa0920958de230"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.489148 4751 generic.go:334] "Generic (PLEG): container finished" podID="d45165ce-0577-4093-b5b2-268035e7540f" containerID="18562bd55ef9ea77c31cbc410d5cb13f30ee95cb01dcb15a076a0f6f573fb1ea" exitCode=0 Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.489208 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l56lv" event={"ID":"d45165ce-0577-4093-b5b2-268035e7540f","Type":"ContainerDied","Data":"18562bd55ef9ea77c31cbc410d5cb13f30ee95cb01dcb15a076a0f6f573fb1ea"} Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.509759 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwlp7\" (UniqueName: \"kubernetes.io/projected/bc55b82d-1e7c-4036-b372-e881a31dcbe5-kube-api-access-fwlp7\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:20 crc kubenswrapper[4751]: I1002 11:14:20.509819 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glvfz\" (UniqueName: \"kubernetes.io/projected/6514a959-a0fd-4226-8676-00c411e7e1fb-kube-api-access-glvfz\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.501821 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"0ceef75769b5c0d062cd9afeda4baacbbccc5158353ce5d63e02e5c9cdfa8253"} Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.502499 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"0e023f57ce90ce1d4352f2a3ef501979446e3b21e20b78c2c4238d74548f96b0"} Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.506972 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53ba58f6-e55e-4bd6-9365-04fe987aec95","Type":"ContainerStarted","Data":"6d97144bed49234b26287ff21ee366df3147aa8f6365fa6e2e55107a601447f6"} Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.507015 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"53ba58f6-e55e-4bd6-9365-04fe987aec95","Type":"ContainerStarted","Data":"a3ebb556d48eeba8091655bb42fa923eb8490455b3462f0b64ce4d9a0860fd81"} Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.917524 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.931584 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.934942 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.999455497 podStartE2EDuration="13.934928953s" podCreationTimestamp="2025-10-02 11:14:08 +0000 UTC" firstStartedPulling="2025-10-02 11:14:09.482067085 +0000 UTC m=+1331.536293545" lastFinishedPulling="2025-10-02 11:14:20.417540551 +0000 UTC m=+1342.471767001" observedRunningTime="2025-10-02 11:14:21.536052641 +0000 UTC m=+1343.590279111" watchObservedRunningTime="2025-10-02 11:14:21.934928953 +0000 UTC m=+1343.989155403" Oct 02 11:14:21 crc kubenswrapper[4751]: I1002 11:14:21.946654 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.031317 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djtnv\" (UniqueName: \"kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv\") pod \"d9944c9c-cead-44e6-b589-788b4ef04cc7\" (UID: \"d9944c9c-cead-44e6-b589-788b4ef04cc7\") " Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.031354 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2svtz\" (UniqueName: \"kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz\") pod \"d45165ce-0577-4093-b5b2-268035e7540f\" (UID: \"d45165ce-0577-4093-b5b2-268035e7540f\") " Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.031493 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zhzl\" (UniqueName: \"kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl\") pod \"16d45636-f990-47b9-923f-f1464ea8c046\" (UID: \"16d45636-f990-47b9-923f-f1464ea8c046\") " Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.038096 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv" (OuterVolumeSpecName: "kube-api-access-djtnv") pod "d9944c9c-cead-44e6-b589-788b4ef04cc7" (UID: "d9944c9c-cead-44e6-b589-788b4ef04cc7"). InnerVolumeSpecName "kube-api-access-djtnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.038138 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz" (OuterVolumeSpecName: "kube-api-access-2svtz") pod "d45165ce-0577-4093-b5b2-268035e7540f" (UID: "d45165ce-0577-4093-b5b2-268035e7540f"). InnerVolumeSpecName "kube-api-access-2svtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.038752 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl" (OuterVolumeSpecName: "kube-api-access-6zhzl") pod "16d45636-f990-47b9-923f-f1464ea8c046" (UID: "16d45636-f990-47b9-923f-f1464ea8c046"). InnerVolumeSpecName "kube-api-access-6zhzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.133714 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zhzl\" (UniqueName: \"kubernetes.io/projected/16d45636-f990-47b9-923f-f1464ea8c046-kube-api-access-6zhzl\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.133769 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djtnv\" (UniqueName: \"kubernetes.io/projected/d9944c9c-cead-44e6-b589-788b4ef04cc7-kube-api-access-djtnv\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.133780 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2svtz\" (UniqueName: \"kubernetes.io/projected/d45165ce-0577-4093-b5b2-268035e7540f-kube-api-access-2svtz\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.517200 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"4a91f03951c96ca0f8d96c61296ac4b12a7bef3011dd8156c7873d983bb5e42b"} Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.517249 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"9e86da59d66d0624fd3221b18f60415544f7f3df8fd343618c3ebf2ee0612174"} Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.518956 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ll6rj" event={"ID":"16d45636-f990-47b9-923f-f1464ea8c046","Type":"ContainerDied","Data":"f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593"} Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.519006 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1079cab00584fa087fabde67126430c905cd46825635bbb0c0b1340dc077593" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.518978 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ll6rj" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.525535 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l56lv" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.525588 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l56lv" event={"ID":"d45165ce-0577-4093-b5b2-268035e7540f","Type":"ContainerDied","Data":"9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65"} Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.525639 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9546d7486d200d437a652083e74e27f7fed51fdd277c2a2a97fb38878fe31a65" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.527161 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pz7qh" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.527241 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pz7qh" event={"ID":"d9944c9c-cead-44e6-b589-788b4ef04cc7","Type":"ContainerDied","Data":"0eccdb14731e19bb2fe35c3ce26e16c2e7b5f16aa14631b5923c2961c9ff1fd4"} Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.527327 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eccdb14731e19bb2fe35c3ce26e16c2e7b5f16aa14631b5923c2961c9ff1fd4" Oct 02 11:14:22 crc kubenswrapper[4751]: I1002 11:14:22.527380 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 11:14:24 crc kubenswrapper[4751]: I1002 11:14:24.558260 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"2fe488e6d0b2aaf2078c989a46ee620fd17747e06be8a4b36b121cd856d88eeb"} Oct 02 11:14:25 crc kubenswrapper[4751]: I1002 11:14:25.587087 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"8041a28214941656dd42e81f142c948c331c80e28c9f39e3b19f6f286d510577"} Oct 02 11:14:25 crc kubenswrapper[4751]: I1002 11:14:25.587473 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"f8f303295e7901904feebd8059e01c00297dafbd8e3e3bc7b5549245370225cc"} Oct 02 11:14:25 crc kubenswrapper[4751]: I1002 11:14:25.587493 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"30ff94fa94bf5f61eda2bb22611073a4cb60d46ed7c2770f871f3129ff489262"} Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.605026 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"3ceb0b8e98429c3bd75e0e58791f57cce81b3049c0ab1ba572c46346eb0c2c1c"} Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.605099 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"a42ba6fe3c710e419432c83b1f16e580edd19cc039980c04ed10a1f8d8671938"} Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.605112 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"b0071ad0-2497-45b8-bddc-9164ee537d18","Type":"ContainerStarted","Data":"af91845d6a017a0a1721832c36ad5c41a1af8b1b526e2849ca31bab28abfefb2"} Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.886937 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-71a7-account-create-nr8bd"] Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.887814 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45165ce-0577-4093-b5b2-268035e7540f" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.887841 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45165ce-0577-4093-b5b2-268035e7540f" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.887868 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a3d249-497a-4ea8-9752-6f21d35f392d" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.887881 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a3d249-497a-4ea8-9752-6f21d35f392d" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.887895 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc55b82d-1e7c-4036-b372-e881a31dcbe5" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.887908 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc55b82d-1e7c-4036-b372-e881a31dcbe5" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.887930 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d45636-f990-47b9-923f-f1464ea8c046" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.887942 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d45636-f990-47b9-923f-f1464ea8c046" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.887998 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9944c9c-cead-44e6-b589-788b4ef04cc7" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888011 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9944c9c-cead-44e6-b589-788b4ef04cc7" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: E1002 11:14:26.888034 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6514a959-a0fd-4226-8676-00c411e7e1fb" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888045 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6514a959-a0fd-4226-8676-00c411e7e1fb" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888363 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="39a3d249-497a-4ea8-9752-6f21d35f392d" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888416 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6514a959-a0fd-4226-8676-00c411e7e1fb" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888449 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc55b82d-1e7c-4036-b372-e881a31dcbe5" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888473 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d45165ce-0577-4093-b5b2-268035e7540f" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888496 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9944c9c-cead-44e6-b589-788b4ef04cc7" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.888514 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="16d45636-f990-47b9-923f-f1464ea8c046" containerName="mariadb-database-create" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.889347 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.891965 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.899403 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-71a7-account-create-nr8bd"] Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.925842 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94cz6\" (UniqueName: \"kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6\") pod \"barbican-71a7-account-create-nr8bd\" (UID: \"9abde181-ba76-4a0e-9346-af67fcd274f2\") " pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.961716 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ea01-account-create-g628k"] Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.967310 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.972857 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 11:14:26 crc kubenswrapper[4751]: I1002 11:14:26.973696 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea01-account-create-g628k"] Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.027733 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqddt\" (UniqueName: \"kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt\") pod \"cinder-ea01-account-create-g628k\" (UID: \"87713887-97a6-40cc-8c3b-b92b64a26112\") " pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.027865 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94cz6\" (UniqueName: \"kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6\") pod \"barbican-71a7-account-create-nr8bd\" (UID: \"9abde181-ba76-4a0e-9346-af67fcd274f2\") " pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.055981 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94cz6\" (UniqueName: \"kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6\") pod \"barbican-71a7-account-create-nr8bd\" (UID: \"9abde181-ba76-4a0e-9346-af67fcd274f2\") " pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.129973 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqddt\" (UniqueName: \"kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt\") pod \"cinder-ea01-account-create-g628k\" (UID: \"87713887-97a6-40cc-8c3b-b92b64a26112\") " pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.147623 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqddt\" (UniqueName: \"kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt\") pod \"cinder-ea01-account-create-g628k\" (UID: \"87713887-97a6-40cc-8c3b-b92b64a26112\") " pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.217587 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.281518 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3e12-account-create-jvktf"] Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.282664 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.284708 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.286669 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.297364 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e12-account-create-jvktf"] Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.333814 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzx2c\" (UniqueName: \"kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c\") pod \"neutron-3e12-account-create-jvktf\" (UID: \"478f1afd-b217-42d6-b69e-3f66c7a893bf\") " pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.436327 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzx2c\" (UniqueName: \"kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c\") pod \"neutron-3e12-account-create-jvktf\" (UID: \"478f1afd-b217-42d6-b69e-3f66c7a893bf\") " pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.460759 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzx2c\" (UniqueName: \"kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c\") pod \"neutron-3e12-account-create-jvktf\" (UID: \"478f1afd-b217-42d6-b69e-3f66c7a893bf\") " pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.651263 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.659124 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.994296734 podStartE2EDuration="56.659103191s" podCreationTimestamp="2025-10-02 11:13:31 +0000 UTC" firstStartedPulling="2025-10-02 11:14:05.484652535 +0000 UTC m=+1327.538878985" lastFinishedPulling="2025-10-02 11:14:24.149458992 +0000 UTC m=+1346.203685442" observedRunningTime="2025-10-02 11:14:27.658093644 +0000 UTC m=+1349.712320124" watchObservedRunningTime="2025-10-02 11:14:27.659103191 +0000 UTC m=+1349.713329661" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.687322 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-71a7-account-create-nr8bd"] Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.747006 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea01-account-create-g628k"] Oct 02 11:14:27 crc kubenswrapper[4751]: W1002 11:14:27.751533 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87713887_97a6_40cc_8c3b_b92b64a26112.slice/crio-d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a WatchSource:0}: Error finding container d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a: Status 404 returned error can't find the container with id d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.943390 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.946201 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.954314 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 11:14:27 crc kubenswrapper[4751]: I1002 11:14:27.957656 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047181 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047253 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047341 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047399 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvvqs\" (UniqueName: \"kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047426 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.047460 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.105402 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e12-account-create-jvktf"] Oct 02 11:14:28 crc kubenswrapper[4751]: W1002 11:14:28.139121 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478f1afd_b217_42d6_b69e_3f66c7a893bf.slice/crio-e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb WatchSource:0}: Error finding container e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb: Status 404 returned error can't find the container with id e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148448 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148520 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148578 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvvqs\" (UniqueName: \"kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148609 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148639 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.148667 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.149781 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.149807 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.150683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.150894 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.151726 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.170000 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvvqs\" (UniqueName: \"kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs\") pod \"dnsmasq-dns-6d5b6d6b67-nrsth\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.263008 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.621738 4751 generic.go:334] "Generic (PLEG): container finished" podID="9abde181-ba76-4a0e-9346-af67fcd274f2" containerID="60bf10fbda72722ef2d9f5aff822f3ba8be4375b9295fef422e24fb517c2ff39" exitCode=0 Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.621931 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-71a7-account-create-nr8bd" event={"ID":"9abde181-ba76-4a0e-9346-af67fcd274f2","Type":"ContainerDied","Data":"60bf10fbda72722ef2d9f5aff822f3ba8be4375b9295fef422e24fb517c2ff39"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.622108 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-71a7-account-create-nr8bd" event={"ID":"9abde181-ba76-4a0e-9346-af67fcd274f2","Type":"ContainerStarted","Data":"ef4eff904ca0db924c28ae9a972cc6a812114203f6135da6e6b2dddbe6d91df2"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.623624 4751 generic.go:334] "Generic (PLEG): container finished" podID="87713887-97a6-40cc-8c3b-b92b64a26112" containerID="887ad8f641f8fe949619cff876c524efe0bb8ecac2d1dcfb83762f34cb0e2605" exitCode=0 Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.623702 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea01-account-create-g628k" event={"ID":"87713887-97a6-40cc-8c3b-b92b64a26112","Type":"ContainerDied","Data":"887ad8f641f8fe949619cff876c524efe0bb8ecac2d1dcfb83762f34cb0e2605"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.623759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea01-account-create-g628k" event={"ID":"87713887-97a6-40cc-8c3b-b92b64a26112","Type":"ContainerStarted","Data":"d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.625380 4751 generic.go:334] "Generic (PLEG): container finished" podID="478f1afd-b217-42d6-b69e-3f66c7a893bf" containerID="3e37fb9ae9ba706d2e7582dbff72766d97ba15cb4e3c95979d821db91651455f" exitCode=0 Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.625452 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e12-account-create-jvktf" event={"ID":"478f1afd-b217-42d6-b69e-3f66c7a893bf","Type":"ContainerDied","Data":"3e37fb9ae9ba706d2e7582dbff72766d97ba15cb4e3c95979d821db91651455f"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.625479 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e12-account-create-jvktf" event={"ID":"478f1afd-b217-42d6-b69e-3f66c7a893bf","Type":"ContainerStarted","Data":"e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb"} Oct 02 11:14:28 crc kubenswrapper[4751]: I1002 11:14:28.701877 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.406159 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-df68-account-create-wjhj7"] Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.407907 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.411404 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.419215 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-df68-account-create-wjhj7"] Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.470689 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r48sg\" (UniqueName: \"kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg\") pod \"keystone-df68-account-create-wjhj7\" (UID: \"bea9cc43-9a45-4a96-9efb-5fb84a515625\") " pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.572562 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r48sg\" (UniqueName: \"kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg\") pod \"keystone-df68-account-create-wjhj7\" (UID: \"bea9cc43-9a45-4a96-9efb-5fb84a515625\") " pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.598480 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r48sg\" (UniqueName: \"kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg\") pod \"keystone-df68-account-create-wjhj7\" (UID: \"bea9cc43-9a45-4a96-9efb-5fb84a515625\") " pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.648856 4751 generic.go:334] "Generic (PLEG): container finished" podID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerID="34093f720742a7a87ebbd4c9148a4c344d3979826b9190411a99bdcb2c8258fe" exitCode=0 Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.648903 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" event={"ID":"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9","Type":"ContainerDied","Data":"34093f720742a7a87ebbd4c9148a4c344d3979826b9190411a99bdcb2c8258fe"} Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.648943 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" event={"ID":"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9","Type":"ContainerStarted","Data":"49c93cd839c8bcfe9ae928b73f6ba0dad8d24486858d001b051f4042ff10f896"} Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.726911 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-dccd-account-create-pj5cq"] Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.729321 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.729772 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.736004 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dccd-account-create-pj5cq"] Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.736235 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.775944 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvv5b\" (UniqueName: \"kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b\") pod \"placement-dccd-account-create-pj5cq\" (UID: \"af8ba996-86ce-4ded-9615-77d93533e33f\") " pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.880040 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvv5b\" (UniqueName: \"kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b\") pod \"placement-dccd-account-create-pj5cq\" (UID: \"af8ba996-86ce-4ded-9615-77d93533e33f\") " pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:29 crc kubenswrapper[4751]: I1002 11:14:29.902018 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvv5b\" (UniqueName: \"kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b\") pod \"placement-dccd-account-create-pj5cq\" (UID: \"af8ba996-86ce-4ded-9615-77d93533e33f\") " pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.000114 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.008048 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-afa2-account-create-x7p58"] Oct 02 11:14:30 crc kubenswrapper[4751]: E1002 11:14:30.008744 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87713887-97a6-40cc-8c3b-b92b64a26112" containerName="mariadb-account-create" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.008770 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="87713887-97a6-40cc-8c3b-b92b64a26112" containerName="mariadb-account-create" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.008980 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="87713887-97a6-40cc-8c3b-b92b64a26112" containerName="mariadb-account-create" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.009923 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.013841 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.025242 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-afa2-account-create-x7p58"] Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.071629 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.082999 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.083293 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqddt\" (UniqueName: \"kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt\") pod \"87713887-97a6-40cc-8c3b-b92b64a26112\" (UID: \"87713887-97a6-40cc-8c3b-b92b64a26112\") " Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.083782 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvtmm\" (UniqueName: \"kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm\") pod \"glance-afa2-account-create-x7p58\" (UID: \"beeb2d4a-d95e-42d7-98bc-13c1b23ea096\") " pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.091101 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt" (OuterVolumeSpecName: "kube-api-access-pqddt") pod "87713887-97a6-40cc-8c3b-b92b64a26112" (UID: "87713887-97a6-40cc-8c3b-b92b64a26112"). InnerVolumeSpecName "kube-api-access-pqddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.115688 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.185245 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzx2c\" (UniqueName: \"kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c\") pod \"478f1afd-b217-42d6-b69e-3f66c7a893bf\" (UID: \"478f1afd-b217-42d6-b69e-3f66c7a893bf\") " Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.185461 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94cz6\" (UniqueName: \"kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6\") pod \"9abde181-ba76-4a0e-9346-af67fcd274f2\" (UID: \"9abde181-ba76-4a0e-9346-af67fcd274f2\") " Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.185789 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvtmm\" (UniqueName: \"kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm\") pod \"glance-afa2-account-create-x7p58\" (UID: \"beeb2d4a-d95e-42d7-98bc-13c1b23ea096\") " pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.185857 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqddt\" (UniqueName: \"kubernetes.io/projected/87713887-97a6-40cc-8c3b-b92b64a26112-kube-api-access-pqddt\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.193361 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6" (OuterVolumeSpecName: "kube-api-access-94cz6") pod "9abde181-ba76-4a0e-9346-af67fcd274f2" (UID: "9abde181-ba76-4a0e-9346-af67fcd274f2"). InnerVolumeSpecName "kube-api-access-94cz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.193525 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c" (OuterVolumeSpecName: "kube-api-access-bzx2c") pod "478f1afd-b217-42d6-b69e-3f66c7a893bf" (UID: "478f1afd-b217-42d6-b69e-3f66c7a893bf"). InnerVolumeSpecName "kube-api-access-bzx2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.202806 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvtmm\" (UniqueName: \"kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm\") pod \"glance-afa2-account-create-x7p58\" (UID: \"beeb2d4a-d95e-42d7-98bc-13c1b23ea096\") " pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.287643 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzx2c\" (UniqueName: \"kubernetes.io/projected/478f1afd-b217-42d6-b69e-3f66c7a893bf-kube-api-access-bzx2c\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.287679 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94cz6\" (UniqueName: \"kubernetes.io/projected/9abde181-ba76-4a0e-9346-af67fcd274f2-kube-api-access-94cz6\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.330750 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.332999 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-df68-account-create-wjhj7"] Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.520847 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-dccd-account-create-pj5cq"] Oct 02 11:14:30 crc kubenswrapper[4751]: W1002 11:14:30.527084 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf8ba996_86ce_4ded_9615_77d93533e33f.slice/crio-4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a WatchSource:0}: Error finding container 4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a: Status 404 returned error can't find the container with id 4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.661102 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dccd-account-create-pj5cq" event={"ID":"af8ba996-86ce-4ded-9615-77d93533e33f","Type":"ContainerStarted","Data":"4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.663299 4751 generic.go:334] "Generic (PLEG): container finished" podID="bea9cc43-9a45-4a96-9efb-5fb84a515625" containerID="5a79a95a25cc83a83077fd6a66d465e5c19b052756122260a117d710e9e80930" exitCode=0 Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.664741 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df68-account-create-wjhj7" event={"ID":"bea9cc43-9a45-4a96-9efb-5fb84a515625","Type":"ContainerDied","Data":"5a79a95a25cc83a83077fd6a66d465e5c19b052756122260a117d710e9e80930"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.664778 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df68-account-create-wjhj7" event={"ID":"bea9cc43-9a45-4a96-9efb-5fb84a515625","Type":"ContainerStarted","Data":"5b9465fa78d846904b568105bc5fd3c06b73e95716af39a18235ffbc0a0d54f8"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.666990 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e12-account-create-jvktf" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.666976 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e12-account-create-jvktf" event={"ID":"478f1afd-b217-42d6-b69e-3f66c7a893bf","Type":"ContainerDied","Data":"e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.667049 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e30c0990104fcd36938ccfd0fcbbdf3b2bfdc5527be0c458c1e9d048becec9cb" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.669117 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-71a7-account-create-nr8bd" event={"ID":"9abde181-ba76-4a0e-9346-af67fcd274f2","Type":"ContainerDied","Data":"ef4eff904ca0db924c28ae9a972cc6a812114203f6135da6e6b2dddbe6d91df2"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.669150 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef4eff904ca0db924c28ae9a972cc6a812114203f6135da6e6b2dddbe6d91df2" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.669196 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-71a7-account-create-nr8bd" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.673227 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea01-account-create-g628k" event={"ID":"87713887-97a6-40cc-8c3b-b92b64a26112","Type":"ContainerDied","Data":"d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.673295 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d950d3160ead41434ca90fbb7bee68bdca57f1c757f98b68347bf9b70f4f841a" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.673340 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea01-account-create-g628k" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.681081 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" event={"ID":"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9","Type":"ContainerStarted","Data":"67429c367231ec46a2c6a8748f4439bbbba9b4d50546c0780641dadcbb65b101"} Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.681394 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.763572 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podStartSLOduration=3.763550938 podStartE2EDuration="3.763550938s" podCreationTimestamp="2025-10-02 11:14:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:30.706621076 +0000 UTC m=+1352.760847696" watchObservedRunningTime="2025-10-02 11:14:30.763550938 +0000 UTC m=+1352.817777388" Oct 02 11:14:30 crc kubenswrapper[4751]: I1002 11:14:30.769501 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-afa2-account-create-x7p58"] Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.507477 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.507566 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.695707 4751 generic.go:334] "Generic (PLEG): container finished" podID="af8ba996-86ce-4ded-9615-77d93533e33f" containerID="472b02f690b09f41d9dcb9a58d7c2cfc58909fdf059e91c3fdf9bf36db484b2f" exitCode=0 Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.695807 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dccd-account-create-pj5cq" event={"ID":"af8ba996-86ce-4ded-9615-77d93533e33f","Type":"ContainerDied","Data":"472b02f690b09f41d9dcb9a58d7c2cfc58909fdf059e91c3fdf9bf36db484b2f"} Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.698763 4751 generic.go:334] "Generic (PLEG): container finished" podID="beeb2d4a-d95e-42d7-98bc-13c1b23ea096" containerID="ee670e75c1498a1a22998d51629e39aca6decdcb1078ff1ad1904f08bf90eb5e" exitCode=0 Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.699149 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-afa2-account-create-x7p58" event={"ID":"beeb2d4a-d95e-42d7-98bc-13c1b23ea096","Type":"ContainerDied","Data":"ee670e75c1498a1a22998d51629e39aca6decdcb1078ff1ad1904f08bf90eb5e"} Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.699244 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-afa2-account-create-x7p58" event={"ID":"beeb2d4a-d95e-42d7-98bc-13c1b23ea096","Type":"ContainerStarted","Data":"0d51502addf8c96bcab9633607b7407433430c542b0ebbe13ba65853e414145a"} Oct 02 11:14:31 crc kubenswrapper[4751]: I1002 11:14:31.997344 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.121618 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r48sg\" (UniqueName: \"kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg\") pod \"bea9cc43-9a45-4a96-9efb-5fb84a515625\" (UID: \"bea9cc43-9a45-4a96-9efb-5fb84a515625\") " Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.131413 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg" (OuterVolumeSpecName: "kube-api-access-r48sg") pod "bea9cc43-9a45-4a96-9efb-5fb84a515625" (UID: "bea9cc43-9a45-4a96-9efb-5fb84a515625"). InnerVolumeSpecName "kube-api-access-r48sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.223279 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r48sg\" (UniqueName: \"kubernetes.io/projected/bea9cc43-9a45-4a96-9efb-5fb84a515625-kube-api-access-r48sg\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.708568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-df68-account-create-wjhj7" event={"ID":"bea9cc43-9a45-4a96-9efb-5fb84a515625","Type":"ContainerDied","Data":"5b9465fa78d846904b568105bc5fd3c06b73e95716af39a18235ffbc0a0d54f8"} Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.708624 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b9465fa78d846904b568105bc5fd3c06b73e95716af39a18235ffbc0a0d54f8" Oct 02 11:14:32 crc kubenswrapper[4751]: I1002 11:14:32.708680 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-df68-account-create-wjhj7" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.122833 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.133697 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.241452 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvv5b\" (UniqueName: \"kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b\") pod \"af8ba996-86ce-4ded-9615-77d93533e33f\" (UID: \"af8ba996-86ce-4ded-9615-77d93533e33f\") " Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.241581 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvtmm\" (UniqueName: \"kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm\") pod \"beeb2d4a-d95e-42d7-98bc-13c1b23ea096\" (UID: \"beeb2d4a-d95e-42d7-98bc-13c1b23ea096\") " Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.249392 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm" (OuterVolumeSpecName: "kube-api-access-kvtmm") pod "beeb2d4a-d95e-42d7-98bc-13c1b23ea096" (UID: "beeb2d4a-d95e-42d7-98bc-13c1b23ea096"). InnerVolumeSpecName "kube-api-access-kvtmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.252924 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b" (OuterVolumeSpecName: "kube-api-access-gvv5b") pod "af8ba996-86ce-4ded-9615-77d93533e33f" (UID: "af8ba996-86ce-4ded-9615-77d93533e33f"). InnerVolumeSpecName "kube-api-access-gvv5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.344054 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvtmm\" (UniqueName: \"kubernetes.io/projected/beeb2d4a-d95e-42d7-98bc-13c1b23ea096-kube-api-access-kvtmm\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.344137 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvv5b\" (UniqueName: \"kubernetes.io/projected/af8ba996-86ce-4ded-9615-77d93533e33f-kube-api-access-gvv5b\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.720729 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-dccd-account-create-pj5cq" event={"ID":"af8ba996-86ce-4ded-9615-77d93533e33f","Type":"ContainerDied","Data":"4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a"} Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.720782 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-dccd-account-create-pj5cq" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.720802 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f1e9fb153bda88089bbd5e9eea9e7563faa21cdd604c5393fa57a3d8ce67c6a" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.722594 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-afa2-account-create-x7p58" event={"ID":"beeb2d4a-d95e-42d7-98bc-13c1b23ea096","Type":"ContainerDied","Data":"0d51502addf8c96bcab9633607b7407433430c542b0ebbe13ba65853e414145a"} Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.722624 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d51502addf8c96bcab9633607b7407433430c542b0ebbe13ba65853e414145a" Oct 02 11:14:33 crc kubenswrapper[4751]: I1002 11:14:33.722678 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-afa2-account-create-x7p58" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.051990 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.813515 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-v4zhn"] Oct 02 11:14:34 crc kubenswrapper[4751]: E1002 11:14:34.814104 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beeb2d4a-d95e-42d7-98bc-13c1b23ea096" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814116 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="beeb2d4a-d95e-42d7-98bc-13c1b23ea096" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: E1002 11:14:34.814129 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea9cc43-9a45-4a96-9efb-5fb84a515625" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814135 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea9cc43-9a45-4a96-9efb-5fb84a515625" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: E1002 11:14:34.814148 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9abde181-ba76-4a0e-9346-af67fcd274f2" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814155 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9abde181-ba76-4a0e-9346-af67fcd274f2" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: E1002 11:14:34.814217 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af8ba996-86ce-4ded-9615-77d93533e33f" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814225 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="af8ba996-86ce-4ded-9615-77d93533e33f" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: E1002 11:14:34.814239 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478f1afd-b217-42d6-b69e-3f66c7a893bf" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814245 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="478f1afd-b217-42d6-b69e-3f66c7a893bf" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814405 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="beeb2d4a-d95e-42d7-98bc-13c1b23ea096" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814426 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="9abde181-ba76-4a0e-9346-af67fcd274f2" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814434 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea9cc43-9a45-4a96-9efb-5fb84a515625" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814441 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="af8ba996-86ce-4ded-9615-77d93533e33f" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814449 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="478f1afd-b217-42d6-b69e-3f66c7a893bf" containerName="mariadb-account-create" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.814945 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.822734 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.823015 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.823236 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gk6gt" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.823817 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.830277 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-v4zhn"] Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.870486 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.870709 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.870753 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnx2v\" (UniqueName: \"kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.972117 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.972258 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnx2v\" (UniqueName: \"kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.972312 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.980340 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.981655 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:34 crc kubenswrapper[4751]: I1002 11:14:34.989949 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnx2v\" (UniqueName: \"kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v\") pod \"keystone-db-sync-v4zhn\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.093880 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-44kgc"] Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.095182 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.097117 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.097388 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xd7tx" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.102909 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-44kgc"] Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.160266 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.175848 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.175913 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.175938 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.176066 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d98pt\" (UniqueName: \"kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.278116 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.278580 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.278606 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.279340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d98pt\" (UniqueName: \"kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.283831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.286257 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.287955 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.296907 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d98pt\" (UniqueName: \"kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt\") pod \"glance-db-sync-44kgc\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.411341 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-44kgc" Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.580252 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-v4zhn"] Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.740163 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v4zhn" event={"ID":"ff6df6e4-d37d-4862-9f23-4932c55784ec","Type":"ContainerStarted","Data":"f229e19a0af18e0bed643aa728546ce811e6f11b156394e66d2f2f3be65c1fae"} Oct 02 11:14:35 crc kubenswrapper[4751]: I1002 11:14:35.962744 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-44kgc"] Oct 02 11:14:35 crc kubenswrapper[4751]: W1002 11:14:35.963843 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d2a342d_289c_41e9_94f7_6ddbc10a6e1c.slice/crio-0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0 WatchSource:0}: Error finding container 0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0: Status 404 returned error can't find the container with id 0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0 Oct 02 11:14:36 crc kubenswrapper[4751]: I1002 11:14:36.753614 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-44kgc" event={"ID":"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c","Type":"ContainerStarted","Data":"0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0"} Oct 02 11:14:38 crc kubenswrapper[4751]: I1002 11:14:38.264587 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:14:38 crc kubenswrapper[4751]: I1002 11:14:38.339044 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:14:38 crc kubenswrapper[4751]: I1002 11:14:38.342052 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="dnsmasq-dns" containerID="cri-o://d985882ebbf19e74598c046f77e673a78cbeb084ac02d84f0e6b16e9c272f4c9" gracePeriod=10 Oct 02 11:14:38 crc kubenswrapper[4751]: I1002 11:14:38.772720 4751 generic.go:334] "Generic (PLEG): container finished" podID="55adacb7-50c1-41b0-9c40-171fdae99382" containerID="d985882ebbf19e74598c046f77e673a78cbeb084ac02d84f0e6b16e9c272f4c9" exitCode=0 Oct 02 11:14:38 crc kubenswrapper[4751]: I1002 11:14:38.772797 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" event={"ID":"55adacb7-50c1-41b0-9c40-171fdae99382","Type":"ContainerDied","Data":"d985882ebbf19e74598c046f77e673a78cbeb084ac02d84f0e6b16e9c272f4c9"} Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.346706 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.383667 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config\") pod \"55adacb7-50c1-41b0-9c40-171fdae99382\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.383810 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc\") pod \"55adacb7-50c1-41b0-9c40-171fdae99382\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.383914 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb\") pod \"55adacb7-50c1-41b0-9c40-171fdae99382\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.384008 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb\") pod \"55adacb7-50c1-41b0-9c40-171fdae99382\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.384045 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpzsw\" (UniqueName: \"kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw\") pod \"55adacb7-50c1-41b0-9c40-171fdae99382\" (UID: \"55adacb7-50c1-41b0-9c40-171fdae99382\") " Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.392119 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw" (OuterVolumeSpecName: "kube-api-access-bpzsw") pod "55adacb7-50c1-41b0-9c40-171fdae99382" (UID: "55adacb7-50c1-41b0-9c40-171fdae99382"). InnerVolumeSpecName "kube-api-access-bpzsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.429002 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55adacb7-50c1-41b0-9c40-171fdae99382" (UID: "55adacb7-50c1-41b0-9c40-171fdae99382"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.441836 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55adacb7-50c1-41b0-9c40-171fdae99382" (UID: "55adacb7-50c1-41b0-9c40-171fdae99382"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.441919 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config" (OuterVolumeSpecName: "config") pod "55adacb7-50c1-41b0-9c40-171fdae99382" (UID: "55adacb7-50c1-41b0-9c40-171fdae99382"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.469838 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55adacb7-50c1-41b0-9c40-171fdae99382" (UID: "55adacb7-50c1-41b0-9c40-171fdae99382"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.487886 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.487922 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.487936 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.487949 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55adacb7-50c1-41b0-9c40-171fdae99382-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.487961 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpzsw\" (UniqueName: \"kubernetes.io/projected/55adacb7-50c1-41b0-9c40-171fdae99382-kube-api-access-bpzsw\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.789873 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v4zhn" event={"ID":"ff6df6e4-d37d-4862-9f23-4932c55784ec","Type":"ContainerStarted","Data":"d137206074fbe87568b54ac62fe82a923f225ef4f3e0ca1c32d3ff70993dfad0"} Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.791582 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" event={"ID":"55adacb7-50c1-41b0-9c40-171fdae99382","Type":"ContainerDied","Data":"c39f89468c39b91c965a9d6b916a52d8a67f1d2844d8c6ca21dd03380ee2bf8c"} Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.791611 4751 scope.go:117] "RemoveContainer" containerID="d985882ebbf19e74598c046f77e673a78cbeb084ac02d84f0e6b16e9c272f4c9" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.791862 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-khsms" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.817391 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-v4zhn" podStartSLOduration=2.319857439 podStartE2EDuration="6.817373464s" podCreationTimestamp="2025-10-02 11:14:34 +0000 UTC" firstStartedPulling="2025-10-02 11:14:35.596880632 +0000 UTC m=+1357.651107092" lastFinishedPulling="2025-10-02 11:14:40.094396667 +0000 UTC m=+1362.148623117" observedRunningTime="2025-10-02 11:14:40.810135451 +0000 UTC m=+1362.864361911" watchObservedRunningTime="2025-10-02 11:14:40.817373464 +0000 UTC m=+1362.871599914" Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.838024 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:14:40 crc kubenswrapper[4751]: I1002 11:14:40.843812 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-khsms"] Oct 02 11:14:41 crc kubenswrapper[4751]: I1002 11:14:41.561550 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" path="/var/lib/kubelet/pods/55adacb7-50c1-41b0-9c40-171fdae99382/volumes" Oct 02 11:14:48 crc kubenswrapper[4751]: I1002 11:14:48.461830 4751 scope.go:117] "RemoveContainer" containerID="0e4d360ae32c0344eb667c53bcd62dc29184ad6259acc91d15468afa34dc6492" Oct 02 11:14:49 crc kubenswrapper[4751]: I1002 11:14:49.884612 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-44kgc" event={"ID":"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c","Type":"ContainerStarted","Data":"52cc443f91d71eb36ec985ec6a87d5dd4b78c443e81219875e00519e8f937532"} Oct 02 11:14:49 crc kubenswrapper[4751]: I1002 11:14:49.907161 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-44kgc" podStartSLOduration=1.999634821 podStartE2EDuration="14.907142003s" podCreationTimestamp="2025-10-02 11:14:35 +0000 UTC" firstStartedPulling="2025-10-02 11:14:35.966241506 +0000 UTC m=+1358.020467956" lastFinishedPulling="2025-10-02 11:14:48.873748678 +0000 UTC m=+1370.927975138" observedRunningTime="2025-10-02 11:14:49.902619652 +0000 UTC m=+1371.956846102" watchObservedRunningTime="2025-10-02 11:14:49.907142003 +0000 UTC m=+1371.961368443" Oct 02 11:14:51 crc kubenswrapper[4751]: I1002 11:14:51.902490 4751 generic.go:334] "Generic (PLEG): container finished" podID="ff6df6e4-d37d-4862-9f23-4932c55784ec" containerID="d137206074fbe87568b54ac62fe82a923f225ef4f3e0ca1c32d3ff70993dfad0" exitCode=0 Oct 02 11:14:51 crc kubenswrapper[4751]: I1002 11:14:51.902860 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v4zhn" event={"ID":"ff6df6e4-d37d-4862-9f23-4932c55784ec","Type":"ContainerDied","Data":"d137206074fbe87568b54ac62fe82a923f225ef4f3e0ca1c32d3ff70993dfad0"} Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.251296 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.400268 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnx2v\" (UniqueName: \"kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v\") pod \"ff6df6e4-d37d-4862-9f23-4932c55784ec\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.400391 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle\") pod \"ff6df6e4-d37d-4862-9f23-4932c55784ec\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.400478 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data\") pod \"ff6df6e4-d37d-4862-9f23-4932c55784ec\" (UID: \"ff6df6e4-d37d-4862-9f23-4932c55784ec\") " Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.460381 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v" (OuterVolumeSpecName: "kube-api-access-vnx2v") pod "ff6df6e4-d37d-4862-9f23-4932c55784ec" (UID: "ff6df6e4-d37d-4862-9f23-4932c55784ec"). InnerVolumeSpecName "kube-api-access-vnx2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.465759 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff6df6e4-d37d-4862-9f23-4932c55784ec" (UID: "ff6df6e4-d37d-4862-9f23-4932c55784ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.492451 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data" (OuterVolumeSpecName: "config-data") pod "ff6df6e4-d37d-4862-9f23-4932c55784ec" (UID: "ff6df6e4-d37d-4862-9f23-4932c55784ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.503198 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.503261 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff6df6e4-d37d-4862-9f23-4932c55784ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.503283 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnx2v\" (UniqueName: \"kubernetes.io/projected/ff6df6e4-d37d-4862-9f23-4932c55784ec-kube-api-access-vnx2v\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.920648 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-v4zhn" event={"ID":"ff6df6e4-d37d-4862-9f23-4932c55784ec","Type":"ContainerDied","Data":"f229e19a0af18e0bed643aa728546ce811e6f11b156394e66d2f2f3be65c1fae"} Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.920746 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-v4zhn" Oct 02 11:14:53 crc kubenswrapper[4751]: I1002 11:14:53.920699 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f229e19a0af18e0bed643aa728546ce811e6f11b156394e66d2f2f3be65c1fae" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196273 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v886j"] Oct 02 11:14:54 crc kubenswrapper[4751]: E1002 11:14:54.196606 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="dnsmasq-dns" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196621 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="dnsmasq-dns" Oct 02 11:14:54 crc kubenswrapper[4751]: E1002 11:14:54.196637 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6df6e4-d37d-4862-9f23-4932c55784ec" containerName="keystone-db-sync" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196645 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6df6e4-d37d-4862-9f23-4932c55784ec" containerName="keystone-db-sync" Oct 02 11:14:54 crc kubenswrapper[4751]: E1002 11:14:54.196666 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="init" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196673 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="init" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196830 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="55adacb7-50c1-41b0-9c40-171fdae99382" containerName="dnsmasq-dns" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.196855 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6df6e4-d37d-4862-9f23-4932c55784ec" containerName="keystone-db-sync" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.197365 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.205927 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.206434 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.210368 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gk6gt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.211986 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.215598 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v886j"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.227432 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.228707 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.238943 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324385 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324506 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324533 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkkhj\" (UniqueName: \"kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324565 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.324610 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.378707 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-28csr"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.380029 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.382873 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ctrq8" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.383055 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.383587 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.385579 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.386881 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.395145 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.395220 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-h68tq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.395473 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.395612 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.401891 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-28csr"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427061 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427352 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427381 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427404 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4zv6\" (UniqueName: \"kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427432 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427465 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427498 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427571 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427607 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427644 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427681 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427706 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427729 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkkhj\" (UniqueName: \"kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427786 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g4n2\" (UniqueName: \"kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427813 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427860 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.427918 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.429618 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.440862 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.455896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.456557 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.457077 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.458030 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.476438 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.476853 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkkhj\" (UniqueName: \"kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj\") pod \"keystone-bootstrap-v886j\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.485446 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.494563 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.509933 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.515453 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v886j" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529261 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529341 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhnnp\" (UniqueName: \"kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529369 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529478 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529501 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529518 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4zv6\" (UniqueName: \"kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529553 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529575 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.529606 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530304 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530383 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530405 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530454 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530475 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530507 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530527 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530551 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g4n2\" (UniqueName: \"kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.530636 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.531415 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.532292 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.535413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.536288 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.536863 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.543715 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.544481 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.556413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.560972 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.568353 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.571859 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g4n2\" (UniqueName: \"kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2\") pod \"cinder-db-sync-28csr\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.587555 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-m2fxt"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.591874 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.595937 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4zv6\" (UniqueName: \"kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6\") pod \"dnsmasq-dns-6f8c45789f-kxpvh\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.601279 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.602805 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.603788 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-stfgf" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.605353 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.609264 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.620622 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m2fxt"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.628394 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631105 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631153 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631188 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631216 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631235 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631251 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631269 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jdbf\" (UniqueName: \"kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631296 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631313 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrnlc\" (UniqueName: \"kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631328 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxp9\" (UniqueName: \"kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631352 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631369 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhnnp\" (UniqueName: \"kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631410 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631430 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631446 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631462 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631483 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631514 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631529 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631545 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.631887 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.632373 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.632599 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.637919 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.660452 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhnnp\" (UniqueName: \"kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp\") pod \"horizon-87cfd899-rmt6w\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.676234 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-jmv2n"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.677655 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.680156 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.681138 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jmv2n"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.685514 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4mw5f" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.697930 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.698561 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.717484 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28csr" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.727984 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.732289 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734291 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734509 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734609 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734672 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734735 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734812 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734878 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.734944 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735001 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735064 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jdbf\" (UniqueName: \"kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735148 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735241 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735306 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrnlc\" (UniqueName: \"kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735366 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxp9\" (UniqueName: \"kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735464 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvt4m\" (UniqueName: \"kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735534 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735593 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735665 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.735732 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.736661 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.737000 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.737897 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.740493 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.745591 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.746642 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.756880 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.756927 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.757347 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.759273 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.762904 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrnlc\" (UniqueName: \"kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.762956 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.762995 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.763507 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data\") pod \"horizon-5b6576bd79-86dlq\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.771946 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jdbf\" (UniqueName: \"kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf\") pod \"ceilometer-0\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.778649 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxp9\" (UniqueName: \"kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9\") pod \"neutron-db-sync-m2fxt\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.793400 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-247rg"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.817590 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-247rg"] Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.817745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.819904 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.820183 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zpwmv" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.820275 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846058 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846264 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846392 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvjkk\" (UniqueName: \"kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846427 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846444 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846471 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b98l4\" (UniqueName: \"kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846515 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846569 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvt4m\" (UniqueName: \"kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846597 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846645 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846767 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846793 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846875 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.846935 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.853546 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.870480 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.878630 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvt4m\" (UniqueName: \"kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m\") pod \"barbican-db-sync-jmv2n\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.925935 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950128 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvjkk\" (UniqueName: \"kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950307 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950326 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b98l4\" (UniqueName: \"kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950343 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950375 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950424 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950453 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950486 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950514 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.950564 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.951004 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.951514 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.951987 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.952899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.955271 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.956012 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.957121 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.961283 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.968953 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.974338 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b98l4\" (UniqueName: \"kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4\") pod \"placement-db-sync-247rg\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " pod="openstack/placement-db-sync-247rg" Oct 02 11:14:54 crc kubenswrapper[4751]: I1002 11:14:54.977795 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvjkk\" (UniqueName: \"kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk\") pod \"dnsmasq-dns-fcfdd6f9f-q25mx\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.019122 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.058745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.087973 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.089284 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.179588 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-247rg" Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.251205 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v886j"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.364983 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.431229 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-28csr"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.441247 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:14:55 crc kubenswrapper[4751]: W1002 11:14:55.444929 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6fac4cc_a2a9_420e_8179_0e835e1ddc9c.slice/crio-3db2d209b005a81d1bed2dcd5da18ae98d0c1ce38f0f68e32cbab3825a37ab01 WatchSource:0}: Error finding container 3db2d209b005a81d1bed2dcd5da18ae98d0c1ce38f0f68e32cbab3825a37ab01: Status 404 returned error can't find the container with id 3db2d209b005a81d1bed2dcd5da18ae98d0c1ce38f0f68e32cbab3825a37ab01 Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.533465 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.663863 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m2fxt"] Oct 02 11:14:55 crc kubenswrapper[4751]: W1002 11:14:55.676283 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6080e003_0583_428a_b263_3badfde0fd37.slice/crio-cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d WatchSource:0}: Error finding container cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d: Status 404 returned error can't find the container with id cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.797826 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-247rg"] Oct 02 11:14:55 crc kubenswrapper[4751]: W1002 11:14:55.823818 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9bb5b1a_3008_4ca4_8c88_dac1b5b22471.slice/crio-a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d WatchSource:0}: Error finding container a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d: Status 404 returned error can't find the container with id a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.826629 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.847204 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.858389 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jmv2n"] Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.954006 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b6576bd79-86dlq" event={"ID":"8a179eac-3e34-41db-9a58-053e34c93874","Type":"ContainerStarted","Data":"9f7fb75d2884da0e9ce5c84f6d71ead94635f20d84850ab3226d4c1ecebb3266"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.958070 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-87cfd899-rmt6w" event={"ID":"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c","Type":"ContainerStarted","Data":"3db2d209b005a81d1bed2dcd5da18ae98d0c1ce38f0f68e32cbab3825a37ab01"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.960488 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-247rg" event={"ID":"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471","Type":"ContainerStarted","Data":"a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.962208 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28csr" event={"ID":"faae59de-da62-46ae-b58e-acb43ada5fec","Type":"ContainerStarted","Data":"dbaa52622de083b9a0172473b0e1bd284c7187db15d0db42c4d895ddeee300e1"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.963971 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" event={"ID":"3e98ade6-6fee-43db-be07-db01e3cd4fec","Type":"ContainerStarted","Data":"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.964006 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" event={"ID":"3e98ade6-6fee-43db-be07-db01e3cd4fec","Type":"ContainerStarted","Data":"7520873ad8ac3513e6f02e731e0c1e404d8342047c234e63ee498e68e23dac0d"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.964227 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" podUID="3e98ade6-6fee-43db-be07-db01e3cd4fec" containerName="init" containerID="cri-o://591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f" gracePeriod=10 Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.967824 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f50a9ca0-4300-4928-8d84-a174dd973a9f","Type":"ContainerStarted","Data":"43e66679553c4834b21b4d579d8d5febd25cbabe442cef0e07e621a5e8e98905"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.973921 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v886j" event={"ID":"521a4ab0-89b5-4284-800e-97306f9d6d54","Type":"ContainerStarted","Data":"42a720db4d767a81d1a1d892ec3f9effbda006ede1bf2910a2d7e2914333993c"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.973966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v886j" event={"ID":"521a4ab0-89b5-4284-800e-97306f9d6d54","Type":"ContainerStarted","Data":"0a341cbaf41dc73c592d8b9a488d37918c18007d568b3a52ee0b075a258291ae"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.977738 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m2fxt" event={"ID":"6080e003-0583-428a-b263-3badfde0fd37","Type":"ContainerStarted","Data":"cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.978949 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" event={"ID":"148c43ee-5684-4168-8590-5128723f150c","Type":"ContainerStarted","Data":"8e87cf0c5f7994844a3c050f7eedd34b53157e27ff44e8a6fcea2f6a1ed73b22"} Oct 02 11:14:55 crc kubenswrapper[4751]: I1002 11:14:55.982616 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmv2n" event={"ID":"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479","Type":"ContainerStarted","Data":"2635bf30974601800c5102c776315a12a9b5f6545f22dd4a332a23170fa0f0a4"} Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.004464 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v886j" podStartSLOduration=2.004446165 podStartE2EDuration="2.004446165s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:56.00125778 +0000 UTC m=+1378.055484230" watchObservedRunningTime="2025-10-02 11:14:56.004446165 +0000 UTC m=+1378.058672615" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.400384 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.436958 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.483281 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.491455 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.512974 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.681572 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ntd\" (UniqueName: \"kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.681685 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.681724 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.681796 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.681858 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.785257 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.785330 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2ntd\" (UniqueName: \"kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.785388 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.785409 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.785482 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.786113 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.786288 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.786750 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.799976 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.802668 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2ntd\" (UniqueName: \"kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd\") pod \"horizon-5dc7bfc975-7gbff\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.856806 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:14:56 crc kubenswrapper[4751]: I1002 11:14:56.993667 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.010782 4751 generic.go:334] "Generic (PLEG): container finished" podID="148c43ee-5684-4168-8590-5128723f150c" containerID="b7662a2245f781b0a60ef019acbee9cf0f655a428ef4cb41935c353d104d7224" exitCode=0 Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.011507 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" event={"ID":"148c43ee-5684-4168-8590-5128723f150c","Type":"ContainerDied","Data":"b7662a2245f781b0a60ef019acbee9cf0f655a428ef4cb41935c353d104d7224"} Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.014997 4751 generic.go:334] "Generic (PLEG): container finished" podID="3e98ade6-6fee-43db-be07-db01e3cd4fec" containerID="591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f" exitCode=0 Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.015042 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" event={"ID":"3e98ade6-6fee-43db-be07-db01e3cd4fec","Type":"ContainerDied","Data":"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f"} Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.015058 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" event={"ID":"3e98ade6-6fee-43db-be07-db01e3cd4fec","Type":"ContainerDied","Data":"7520873ad8ac3513e6f02e731e0c1e404d8342047c234e63ee498e68e23dac0d"} Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.015073 4751 scope.go:117] "RemoveContainer" containerID="591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.015159 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-kxpvh" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.018292 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m2fxt" event={"ID":"6080e003-0583-428a-b263-3badfde0fd37","Type":"ContainerStarted","Data":"a728fae2f1a380c15e5618cdda55c64f9b462f8bb159e26d29d607b390c303a3"} Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.076361 4751 scope.go:117] "RemoveContainer" containerID="591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f" Oct 02 11:14:57 crc kubenswrapper[4751]: E1002 11:14:57.083158 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f\": container with ID starting with 591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f not found: ID does not exist" containerID="591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.083247 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f"} err="failed to get container status \"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f\": rpc error: code = NotFound desc = could not find container \"591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f\": container with ID starting with 591467bdf061dfcf0e55d0a9b06f7357eb646c1359bbc2a7ef33fed46a1b9a9f not found: ID does not exist" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.089832 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-m2fxt" podStartSLOduration=3.089811729 podStartE2EDuration="3.089811729s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:57.052750068 +0000 UTC m=+1379.106976508" watchObservedRunningTime="2025-10-02 11:14:57.089811729 +0000 UTC m=+1379.144038179" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.194924 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.195073 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.195118 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zv6\" (UniqueName: \"kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.195238 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.195265 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.195290 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config\") pod \"3e98ade6-6fee-43db-be07-db01e3cd4fec\" (UID: \"3e98ade6-6fee-43db-be07-db01e3cd4fec\") " Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.210127 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6" (OuterVolumeSpecName: "kube-api-access-x4zv6") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "kube-api-access-x4zv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.231091 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.233786 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.237791 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config" (OuterVolumeSpecName: "config") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.238626 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.265358 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e98ade6-6fee-43db-be07-db01e3cd4fec" (UID: "3e98ade6-6fee-43db-be07-db01e3cd4fec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299071 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299110 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zv6\" (UniqueName: \"kubernetes.io/projected/3e98ade6-6fee-43db-be07-db01e3cd4fec-kube-api-access-x4zv6\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299122 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299132 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299144 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.299152 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e98ade6-6fee-43db-be07-db01e3cd4fec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.372850 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.378508 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-kxpvh"] Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.406637 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:14:57 crc kubenswrapper[4751]: I1002 11:14:57.577131 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e98ade6-6fee-43db-be07-db01e3cd4fec" path="/var/lib/kubelet/pods/3e98ade6-6fee-43db-be07-db01e3cd4fec/volumes" Oct 02 11:14:58 crc kubenswrapper[4751]: I1002 11:14:58.032057 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc7bfc975-7gbff" event={"ID":"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af","Type":"ContainerStarted","Data":"4cd3d6756f1f3480fb9c27c4380ed0d74335d884a3c5d3e078bdf4a610b67e55"} Oct 02 11:14:59 crc kubenswrapper[4751]: I1002 11:14:59.070958 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" event={"ID":"148c43ee-5684-4168-8590-5128723f150c","Type":"ContainerStarted","Data":"7fa9e8e4ed0c632999b45ce22cf369ceb740723c3d095862ba913072af850446"} Oct 02 11:14:59 crc kubenswrapper[4751]: I1002 11:14:59.071214 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:14:59 crc kubenswrapper[4751]: I1002 11:14:59.095929 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" podStartSLOduration=5.095910245 podStartE2EDuration="5.095910245s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:14:59.093916522 +0000 UTC m=+1381.148142982" watchObservedRunningTime="2025-10-02 11:14:59.095910245 +0000 UTC m=+1381.150136705" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.130412 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p"] Oct 02 11:15:00 crc kubenswrapper[4751]: E1002 11:15:00.131071 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e98ade6-6fee-43db-be07-db01e3cd4fec" containerName="init" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.131083 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e98ade6-6fee-43db-be07-db01e3cd4fec" containerName="init" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.131307 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e98ade6-6fee-43db-be07-db01e3cd4fec" containerName="init" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.131969 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.134001 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.134286 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.153793 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7dh4\" (UniqueName: \"kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.153973 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.154043 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.162562 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p"] Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.255701 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.255789 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.255841 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7dh4\" (UniqueName: \"kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.256715 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.262116 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.274408 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7dh4\" (UniqueName: \"kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4\") pod \"collect-profiles-29323395-r4q4p\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.472776 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:00 crc kubenswrapper[4751]: I1002 11:15:00.949011 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p"] Oct 02 11:15:00 crc kubenswrapper[4751]: W1002 11:15:00.972025 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1bfc847_f75b_49bf_9249_b79d5a6139bd.slice/crio-ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff WatchSource:0}: Error finding container ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff: Status 404 returned error can't find the container with id ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.105463 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" event={"ID":"b1bfc847-f75b-49bf-9249-b79d5a6139bd","Type":"ContainerStarted","Data":"ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff"} Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.508795 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.509214 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.509273 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.510026 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:15:01 crc kubenswrapper[4751]: I1002 11:15:01.510080 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9" gracePeriod=600 Oct 02 11:15:02 crc kubenswrapper[4751]: I1002 11:15:02.146980 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9" exitCode=0 Oct 02 11:15:02 crc kubenswrapper[4751]: I1002 11:15:02.147126 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9"} Oct 02 11:15:02 crc kubenswrapper[4751]: I1002 11:15:02.147163 4751 scope.go:117] "RemoveContainer" containerID="2d0afaca4fe6d04104482aa8fe54d6c2148be96ff466147d1905b73da2007182" Oct 02 11:15:02 crc kubenswrapper[4751]: I1002 11:15:02.150144 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" event={"ID":"b1bfc847-f75b-49bf-9249-b79d5a6139bd","Type":"ContainerStarted","Data":"3daa9e98d927dae2c70cb8883cedec1aa51cbe28e65b93038e46498c5080de24"} Oct 02 11:15:02 crc kubenswrapper[4751]: I1002 11:15:02.983094 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.003995 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.006924 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.015523 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.021074 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.062797 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.097749 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cf9f465b8-rf9bv"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.099528 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104795 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfbx9\" (UniqueName: \"kubernetes.io/projected/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-kube-api-access-qfbx9\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104890 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104913 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-logs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104940 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-tls-certs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104962 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-secret-key\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.104989 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105022 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj5vq\" (UniqueName: \"kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105048 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-config-data\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105091 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105198 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cf9f465b8-rf9bv"] Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105153 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105378 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-scripts\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.105423 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-combined-ca-bundle\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.178652 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" podStartSLOduration=3.178631164 podStartE2EDuration="3.178631164s" podCreationTimestamp="2025-10-02 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:03.173990349 +0000 UTC m=+1385.228216809" watchObservedRunningTime="2025-10-02 11:15:03.178631164 +0000 UTC m=+1385.232857614" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.206517 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.206563 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-scripts\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.206592 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-combined-ca-bundle\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.206633 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfbx9\" (UniqueName: \"kubernetes.io/projected/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-kube-api-access-qfbx9\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.206649 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207086 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207259 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-logs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207304 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-tls-certs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207325 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-secret-key\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207354 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207382 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj5vq\" (UniqueName: \"kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207402 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207421 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-config-data\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207438 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207825 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-logs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.207989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-scripts\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.208430 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.208741 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.209417 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-config-data\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.212534 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.212839 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-combined-ca-bundle\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.213384 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.213904 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.214986 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-tls-certs\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.218108 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-horizon-secret-key\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.223241 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj5vq\" (UniqueName: \"kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq\") pod \"horizon-758c87b4db-9d6vw\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.231144 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfbx9\" (UniqueName: \"kubernetes.io/projected/3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd-kube-api-access-qfbx9\") pod \"horizon-6cf9f465b8-rf9bv\" (UID: \"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd\") " pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.329765 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:03 crc kubenswrapper[4751]: I1002 11:15:03.424374 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:05 crc kubenswrapper[4751]: I1002 11:15:05.090249 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:15:05 crc kubenswrapper[4751]: I1002 11:15:05.150132 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:15:05 crc kubenswrapper[4751]: I1002 11:15:05.150435 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" containerID="cri-o://67429c367231ec46a2c6a8748f4439bbbba9b4d50546c0780641dadcbb65b101" gracePeriod=10 Oct 02 11:15:06 crc kubenswrapper[4751]: I1002 11:15:06.210391 4751 generic.go:334] "Generic (PLEG): container finished" podID="b1bfc847-f75b-49bf-9249-b79d5a6139bd" containerID="3daa9e98d927dae2c70cb8883cedec1aa51cbe28e65b93038e46498c5080de24" exitCode=0 Oct 02 11:15:06 crc kubenswrapper[4751]: I1002 11:15:06.210597 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" event={"ID":"b1bfc847-f75b-49bf-9249-b79d5a6139bd","Type":"ContainerDied","Data":"3daa9e98d927dae2c70cb8883cedec1aa51cbe28e65b93038e46498c5080de24"} Oct 02 11:15:08 crc kubenswrapper[4751]: I1002 11:15:08.231562 4751 generic.go:334] "Generic (PLEG): container finished" podID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerID="67429c367231ec46a2c6a8748f4439bbbba9b4d50546c0780641dadcbb65b101" exitCode=0 Oct 02 11:15:08 crc kubenswrapper[4751]: I1002 11:15:08.231704 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" event={"ID":"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9","Type":"ContainerDied","Data":"67429c367231ec46a2c6a8748f4439bbbba9b4d50546c0780641dadcbb65b101"} Oct 02 11:15:08 crc kubenswrapper[4751]: I1002 11:15:08.263778 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 02 11:15:13 crc kubenswrapper[4751]: I1002 11:15:13.264004 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 02 11:15:14 crc kubenswrapper[4751]: E1002 11:15:14.506052 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 02 11:15:14 crc kubenswrapper[4751]: E1002 11:15:14.506281 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65h567hcdh85h5c6h5fbh685h5dch54dh64h5c8hc4h5b6h566h688h665h64fh549h5dch565hf4h67ch5d4h8dhf6h5d6h5cdh59dh7hfch5ch549q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9jdbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f50a9ca0-4300-4928-8d84-a174dd973a9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:18 crc kubenswrapper[4751]: I1002 11:15:18.264786 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 02 11:15:18 crc kubenswrapper[4751]: I1002 11:15:18.265599 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.470537 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.640896 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume\") pod \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.640949 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume\") pod \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.641241 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7dh4\" (UniqueName: \"kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4\") pod \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\" (UID: \"b1bfc847-f75b-49bf-9249-b79d5a6139bd\") " Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.642588 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "b1bfc847-f75b-49bf-9249-b79d5a6139bd" (UID: "b1bfc847-f75b-49bf-9249-b79d5a6139bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.646820 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4" (OuterVolumeSpecName: "kube-api-access-n7dh4") pod "b1bfc847-f75b-49bf-9249-b79d5a6139bd" (UID: "b1bfc847-f75b-49bf-9249-b79d5a6139bd"). InnerVolumeSpecName "kube-api-access-n7dh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.648048 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b1bfc847-f75b-49bf-9249-b79d5a6139bd" (UID: "b1bfc847-f75b-49bf-9249-b79d5a6139bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.743888 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7dh4\" (UniqueName: \"kubernetes.io/projected/b1bfc847-f75b-49bf-9249-b79d5a6139bd-kube-api-access-n7dh4\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.744114 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1bfc847-f75b-49bf-9249-b79d5a6139bd-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:21 crc kubenswrapper[4751]: I1002 11:15:21.744249 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1bfc847-f75b-49bf-9249-b79d5a6139bd-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:22 crc kubenswrapper[4751]: I1002 11:15:22.383048 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" event={"ID":"b1bfc847-f75b-49bf-9249-b79d5a6139bd","Type":"ContainerDied","Data":"ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff"} Oct 02 11:15:22 crc kubenswrapper[4751]: I1002 11:15:22.383101 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddf6ed9f6f9f66fd5f80839e9862ed1cdd3adec22871343c66d6b4dcbfb45cff" Oct 02 11:15:22 crc kubenswrapper[4751]: I1002 11:15:22.383136 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p" Oct 02 11:15:23 crc kubenswrapper[4751]: I1002 11:15:23.264144 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 02 11:15:24 crc kubenswrapper[4751]: E1002 11:15:24.673766 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 11:15:24 crc kubenswrapper[4751]: E1002 11:15:24.674013 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9fh565h576h67chdfh689h5cfhf4hbbh88hc9h59bh5f8hd6h74hd8h686hdbhf8h696h595h554h7bh57h5fbh74hb5h95hdhb6h5bdh5c7q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mrnlc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5b6576bd79-86dlq_openstack(8a179eac-3e34-41db-9a58-053e34c93874): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:24 crc kubenswrapper[4751]: E1002 11:15:24.679955 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5b6576bd79-86dlq" podUID="8a179eac-3e34-41db-9a58-053e34c93874" Oct 02 11:15:26 crc kubenswrapper[4751]: E1002 11:15:26.969889 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 11:15:26 crc kubenswrapper[4751]: E1002 11:15:26.970627 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dvt4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-jmv2n_openstack(d0a15cb7-4b3b-4b25-8b11-1b8f14b06479): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:26 crc kubenswrapper[4751]: E1002 11:15:26.971750 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-jmv2n" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" Oct 02 11:15:27 crc kubenswrapper[4751]: E1002 11:15:27.426588 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-jmv2n" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" Oct 02 11:15:28 crc kubenswrapper[4751]: I1002 11:15:28.435857 4751 generic.go:334] "Generic (PLEG): container finished" podID="521a4ab0-89b5-4284-800e-97306f9d6d54" containerID="42a720db4d767a81d1a1d892ec3f9effbda006ede1bf2910a2d7e2914333993c" exitCode=0 Oct 02 11:15:28 crc kubenswrapper[4751]: I1002 11:15:28.435910 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v886j" event={"ID":"521a4ab0-89b5-4284-800e-97306f9d6d54","Type":"ContainerDied","Data":"42a720db4d767a81d1a1d892ec3f9effbda006ede1bf2910a2d7e2914333993c"} Oct 02 11:15:33 crc kubenswrapper[4751]: I1002 11:15:33.264471 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 02 11:15:38 crc kubenswrapper[4751]: I1002 11:15:38.265692 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 02 11:15:41 crc kubenswrapper[4751]: E1002 11:15:41.913415 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 11:15:41 crc kubenswrapper[4751]: E1002 11:15:41.914127 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65bh4h5f5h7dhb5h9ch66bh57dh68hch55bh576h64chffh66dh575h86h5d5h66h56bh584h549h64ch5cch566h66h9dhb9h96h7h555h5b9q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-87cfd899-rmt6w_openstack(d6fac4cc-a2a9-420e-8179-0e835e1ddc9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:41 crc kubenswrapper[4751]: E1002 11:15:41.916811 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-87cfd899-rmt6w" podUID="d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.267548 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.361480 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.361892 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b98l4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-247rg_openstack(c9bb5b1a-3008-4ca4-8c88-dac1b5b22471): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.363110 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-247rg" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.428715 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.436154 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v886j" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.443701 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.519844 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.520037 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55dh5d4h699h5dh66bh87h684h5b8h5fdh58fh84h5bch674h68dh8bh546h9h648h5bh65bh67fh84h4h658h56bh5b8h6chddh64bh645h58bhf6q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w2ntd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5dc7bfc975-7gbff_openstack(4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.522199 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5dc7bfc975-7gbff" podUID="4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.539868 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts\") pod \"8a179eac-3e34-41db-9a58-053e34c93874\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.539948 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key\") pod \"8a179eac-3e34-41db-9a58-053e34c93874\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.539978 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540036 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540098 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540159 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data\") pod \"8a179eac-3e34-41db-9a58-053e34c93874\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540258 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs\") pod \"8a179eac-3e34-41db-9a58-053e34c93874\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540301 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540452 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540499 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts" (OuterVolumeSpecName: "scripts") pod "8a179eac-3e34-41db-9a58-053e34c93874" (UID: "8a179eac-3e34-41db-9a58-053e34c93874"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540517 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrnlc\" (UniqueName: \"kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc\") pod \"8a179eac-3e34-41db-9a58-053e34c93874\" (UID: \"8a179eac-3e34-41db-9a58-053e34c93874\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540543 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540611 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540662 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540708 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs" (OuterVolumeSpecName: "logs") pod "8a179eac-3e34-41db-9a58-053e34c93874" (UID: "8a179eac-3e34-41db-9a58-053e34c93874"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.540719 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.541250 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.541310 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkkhj\" (UniqueName: \"kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj\") pod \"521a4ab0-89b5-4284-800e-97306f9d6d54\" (UID: \"521a4ab0-89b5-4284-800e-97306f9d6d54\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.541366 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvvqs\" (UniqueName: \"kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs\") pod \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\" (UID: \"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9\") " Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.541964 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a179eac-3e34-41db-9a58-053e34c93874-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.541981 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.542474 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data" (OuterVolumeSpecName: "config-data") pod "8a179eac-3e34-41db-9a58-053e34c93874" (UID: "8a179eac-3e34-41db-9a58-053e34c93874"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.546985 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc" (OuterVolumeSpecName: "kube-api-access-mrnlc") pod "8a179eac-3e34-41db-9a58-053e34c93874" (UID: "8a179eac-3e34-41db-9a58-053e34c93874"). InnerVolumeSpecName "kube-api-access-mrnlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.547023 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.547100 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8a179eac-3e34-41db-9a58-053e34c93874" (UID: "8a179eac-3e34-41db-9a58-053e34c93874"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.548422 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts" (OuterVolumeSpecName: "scripts") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.548473 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.550693 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs" (OuterVolumeSpecName: "kube-api-access-qvvqs") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "kube-api-access-qvvqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.551441 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj" (OuterVolumeSpecName: "kube-api-access-hkkhj") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "kube-api-access-hkkhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.572298 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v886j" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.574470 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b6576bd79-86dlq" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.579644 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.580151 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" Oct 02 11:15:43 crc kubenswrapper[4751]: E1002 11:15:43.581983 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-247rg" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.595355 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data" (OuterVolumeSpecName: "config-data") pod "521a4ab0-89b5-4284-800e-97306f9d6d54" (UID: "521a4ab0-89b5-4284-800e-97306f9d6d54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.614510 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.615341 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config" (OuterVolumeSpecName: "config") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.619807 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.630868 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.637577 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" (UID: "57a113ee-7eb3-4654-a0ad-48e5f69b6ed9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644298 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644344 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644358 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a179eac-3e34-41db-9a58-053e34c93874-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644370 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644382 4751 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644394 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrnlc\" (UniqueName: \"kubernetes.io/projected/8a179eac-3e34-41db-9a58-053e34c93874-kube-api-access-mrnlc\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644405 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644416 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644426 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644437 4751 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644448 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644460 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkkhj\" (UniqueName: \"kubernetes.io/projected/521a4ab0-89b5-4284-800e-97306f9d6d54-kube-api-access-hkkhj\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644472 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvvqs\" (UniqueName: \"kubernetes.io/projected/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9-kube-api-access-qvvqs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644484 4751 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a179eac-3e34-41db-9a58-053e34c93874-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.644496 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/521a4ab0-89b5-4284-800e-97306f9d6d54-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.703446 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v886j" event={"ID":"521a4ab0-89b5-4284-800e-97306f9d6d54","Type":"ContainerDied","Data":"0a341cbaf41dc73c592d8b9a488d37918c18007d568b3a52ee0b075a258291ae"} Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.703498 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a341cbaf41dc73c592d8b9a488d37918c18007d568b3a52ee0b075a258291ae" Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.703544 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b6576bd79-86dlq" event={"ID":"8a179eac-3e34-41db-9a58-053e34c93874","Type":"ContainerDied","Data":"9f7fb75d2884da0e9ce5c84f6d71ead94635f20d84850ab3226d4c1ecebb3266"} Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.703567 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" event={"ID":"57a113ee-7eb3-4654-a0ad-48e5f69b6ed9","Type":"ContainerDied","Data":"49c93cd839c8bcfe9ae928b73f6ba0dad8d24486858d001b051f4042ff10f896"} Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.752442 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.762126 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b6576bd79-86dlq"] Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.912462 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:15:43 crc kubenswrapper[4751]: I1002 11:15:43.918511 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrsth"] Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.609507 4751 scope.go:117] "RemoveContainer" containerID="67429c367231ec46a2c6a8748f4439bbbba9b4d50546c0780641dadcbb65b101" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.646602 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v886j"] Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.653766 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v886j"] Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.679959 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.680187 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7g4n2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-28csr_openstack(faae59de-da62-46ae-b58e-acb43ada5fec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.682325 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-28csr" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.702923 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.712677 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.736845 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fbknj"] Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.744363 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744392 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.744428 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="init" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744438 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="init" Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.744455 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bfc847-f75b-49bf-9249-b79d5a6139bd" containerName="collect-profiles" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744464 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bfc847-f75b-49bf-9249-b79d5a6139bd" containerName="collect-profiles" Oct 02 11:15:44 crc kubenswrapper[4751]: E1002 11:15:44.744476 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521a4ab0-89b5-4284-800e-97306f9d6d54" containerName="keystone-bootstrap" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744486 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="521a4ab0-89b5-4284-800e-97306f9d6d54" containerName="keystone-bootstrap" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744686 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1bfc847-f75b-49bf-9249-b79d5a6139bd" containerName="collect-profiles" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744723 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="521a4ab0-89b5-4284-800e-97306f9d6d54" containerName="keystone-bootstrap" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.744736 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.745625 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.748988 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gk6gt" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.749131 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.749257 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.749788 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.758814 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fbknj"] Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.763988 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2ntd\" (UniqueName: \"kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd\") pod \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.764083 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key\") pod \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.764143 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts\") pod \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.764567 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts" (OuterVolumeSpecName: "scripts") pod "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" (UID: "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.764755 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhnnp\" (UniqueName: \"kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp\") pod \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.765887 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data" (OuterVolumeSpecName: "config-data") pod "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" (UID: "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.764792 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data\") pod \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.766680 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs\") pod \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.766707 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs\") pod \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.766747 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data\") pod \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.766772 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts\") pod \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\" (UID: \"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.766798 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key\") pod \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\" (UID: \"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af\") " Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767055 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767220 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767257 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5xv2\" (UniqueName: \"kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767316 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767333 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs" (OuterVolumeSpecName: "logs") pod "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" (UID: "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.767695 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts" (OuterVolumeSpecName: "scripts") pod "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" (UID: "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768357 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768473 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768356 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs" (OuterVolumeSpecName: "logs") pod "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" (UID: "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768676 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768697 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768708 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768719 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.768758 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.770426 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data" (OuterVolumeSpecName: "config-data") pod "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" (UID: "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.771102 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" (UID: "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.771119 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd" (OuterVolumeSpecName: "kube-api-access-w2ntd") pod "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" (UID: "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af"). InnerVolumeSpecName "kube-api-access-w2ntd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.771678 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" (UID: "4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.774853 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp" (OuterVolumeSpecName: "kube-api-access-lhnnp") pod "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" (UID: "d6fac4cc-a2a9-420e-8179-0e835e1ddc9c"). InnerVolumeSpecName "kube-api-access-lhnnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870065 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870236 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5xv2\" (UniqueName: \"kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870543 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870621 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870731 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhnnp\" (UniqueName: \"kubernetes.io/projected/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-kube-api-access-lhnnp\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870747 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870761 4751 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870775 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2ntd\" (UniqueName: \"kubernetes.io/projected/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af-kube-api-access-w2ntd\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.870786 4751 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.873723 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.873968 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.874112 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.874874 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.876847 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:44 crc kubenswrapper[4751]: I1002 11:15:44.888505 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5xv2\" (UniqueName: \"kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2\") pod \"keystone-bootstrap-fbknj\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.075806 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:45 crc kubenswrapper[4751]: E1002 11:15:45.133600 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified" Oct 02 11:15:45 crc kubenswrapper[4751]: E1002 11:15:45.133802 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-notification-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65h567hcdh85h5c6h5fbh685h5dch54dh64h5c8hc4h5b6h566h688h665h64fh549h5dch565hf4h67ch5d4h8dhf6h5d6h5cdh59dh7hfch5ch549q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-notification-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9jdbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f50a9ca0-4300-4928-8d84-a174dd973a9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.204811 4751 scope.go:117] "RemoveContainer" containerID="34093f720742a7a87ebbd4c9148a4c344d3979826b9190411a99bdcb2c8258fe" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.563560 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521a4ab0-89b5-4284-800e-97306f9d6d54" path="/var/lib/kubelet/pods/521a4ab0-89b5-4284-800e-97306f9d6d54/volumes" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.564753 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" path="/var/lib/kubelet/pods/57a113ee-7eb3-4654-a0ad-48e5f69b6ed9/volumes" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.565439 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a179eac-3e34-41db-9a58-053e34c93874" path="/var/lib/kubelet/pods/8a179eac-3e34-41db-9a58-053e34c93874/volumes" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.607489 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc7bfc975-7gbff" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.607495 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc7bfc975-7gbff" event={"ID":"4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af","Type":"ContainerDied","Data":"4cd3d6756f1f3480fb9c27c4380ed0d74335d884a3c5d3e078bdf4a610b67e55"} Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.616559 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168"} Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.619728 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-87cfd899-rmt6w" event={"ID":"d6fac4cc-a2a9-420e-8179-0e835e1ddc9c","Type":"ContainerDied","Data":"3db2d209b005a81d1bed2dcd5da18ae98d0c1ce38f0f68e32cbab3825a37ab01"} Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.619754 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-87cfd899-rmt6w" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.621429 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cf9f465b8-rf9bv"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.637300 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmv2n" event={"ID":"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479","Type":"ContainerStarted","Data":"119e0a5ecc875d6dce924f32efe9df3e575a54fe719c0861879bebe767dff91f"} Oct 02 11:15:45 crc kubenswrapper[4751]: E1002 11:15:45.639413 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-28csr" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" Oct 02 11:15:45 crc kubenswrapper[4751]: W1002 11:15:45.640125 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b70e752_8ff4_47eb_ba3f_ae41af3fb8dd.slice/crio-8afa369a7446ee5aff84824ba4b5b8a07a6b8b1a5b3852e7a1b0afb0bb5ffbd3 WatchSource:0}: Error finding container 8afa369a7446ee5aff84824ba4b5b8a07a6b8b1a5b3852e7a1b0afb0bb5ffbd3: Status 404 returned error can't find the container with id 8afa369a7446ee5aff84824ba4b5b8a07a6b8b1a5b3852e7a1b0afb0bb5ffbd3 Oct 02 11:15:45 crc kubenswrapper[4751]: W1002 11:15:45.698715 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61916e69_5770_4e34_acae_3b8a551df701.slice/crio-cf3033c6202b89b505b55e5ea2f831594eda5eaf8f1ddca02e35058cf984b123 WatchSource:0}: Error finding container cf3033c6202b89b505b55e5ea2f831594eda5eaf8f1ddca02e35058cf984b123: Status 404 returned error can't find the container with id cf3033c6202b89b505b55e5ea2f831594eda5eaf8f1ddca02e35058cf984b123 Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.703749 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.711142 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-jmv2n" podStartSLOduration=2.300485088 podStartE2EDuration="51.711121023s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="2025-10-02 11:14:55.864082222 +0000 UTC m=+1377.918308672" lastFinishedPulling="2025-10-02 11:15:45.274718157 +0000 UTC m=+1427.328944607" observedRunningTime="2025-10-02 11:15:45.68070848 +0000 UTC m=+1427.734934950" watchObservedRunningTime="2025-10-02 11:15:45.711121023 +0000 UTC m=+1427.765347473" Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.763225 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fbknj"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.780562 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.793881 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5dc7bfc975-7gbff"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.815716 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:15:45 crc kubenswrapper[4751]: I1002 11:15:45.820407 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-87cfd899-rmt6w"] Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.647187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf9f465b8-rf9bv" event={"ID":"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd","Type":"ContainerStarted","Data":"583468a791f62e90c9f9f629371fffa494f42c8f3f3e85447ed7c3c35785bac2"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.647776 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf9f465b8-rf9bv" event={"ID":"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd","Type":"ContainerStarted","Data":"e4ec4fcfdc6a2c5cd443cc5aa3053fad5fdf31aa9fff16d37999a7c3a90286c8"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.647791 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf9f465b8-rf9bv" event={"ID":"3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd","Type":"ContainerStarted","Data":"8afa369a7446ee5aff84824ba4b5b8a07a6b8b1a5b3852e7a1b0afb0bb5ffbd3"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.652306 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerStarted","Data":"d56322dba3fe24647b339aa90dbabe8be5b32654c39498af79c501dcd44d87ac"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.652372 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerStarted","Data":"f351c5cafadb1cd76de3df169a58c548a4b6d6e4f9c992ba4f53d1b041ba6cdf"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.652391 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerStarted","Data":"cf3033c6202b89b505b55e5ea2f831594eda5eaf8f1ddca02e35058cf984b123"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.654442 4751 generic.go:334] "Generic (PLEG): container finished" podID="4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" containerID="52cc443f91d71eb36ec985ec6a87d5dd4b78c443e81219875e00519e8f937532" exitCode=0 Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.654526 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-44kgc" event={"ID":"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c","Type":"ContainerDied","Data":"52cc443f91d71eb36ec985ec6a87d5dd4b78c443e81219875e00519e8f937532"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.656904 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fbknj" event={"ID":"f251a17b-99b0-4656-919b-1d03af5331a4","Type":"ContainerStarted","Data":"50836859af1ede94a8266a964e92ff2894489e3be518f679a8fe934fb90b06e4"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.656931 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fbknj" event={"ID":"f251a17b-99b0-4656-919b-1d03af5331a4","Type":"ContainerStarted","Data":"08473cf3dbf7564d128107290e18898600fb8eaab40c9a11ac7132744d483774"} Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.676872 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cf9f465b8-rf9bv" podStartSLOduration=43.172303011 podStartE2EDuration="43.676852969s" podCreationTimestamp="2025-10-02 11:15:03 +0000 UTC" firstStartedPulling="2025-10-02 11:15:45.643632909 +0000 UTC m=+1427.697859369" lastFinishedPulling="2025-10-02 11:15:46.148182877 +0000 UTC m=+1428.202409327" observedRunningTime="2025-10-02 11:15:46.670636003 +0000 UTC m=+1428.724862473" watchObservedRunningTime="2025-10-02 11:15:46.676852969 +0000 UTC m=+1428.731079429" Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.721141 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-758c87b4db-9d6vw" podStartSLOduration=44.153710845 podStartE2EDuration="44.721117083s" podCreationTimestamp="2025-10-02 11:15:02 +0000 UTC" firstStartedPulling="2025-10-02 11:15:45.701903507 +0000 UTC m=+1427.756129957" lastFinishedPulling="2025-10-02 11:15:46.269309745 +0000 UTC m=+1428.323536195" observedRunningTime="2025-10-02 11:15:46.694494521 +0000 UTC m=+1428.748720971" watchObservedRunningTime="2025-10-02 11:15:46.721117083 +0000 UTC m=+1428.775343563" Oct 02 11:15:46 crc kubenswrapper[4751]: I1002 11:15:46.721560 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fbknj" podStartSLOduration=2.721554444 podStartE2EDuration="2.721554444s" podCreationTimestamp="2025-10-02 11:15:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:46.710819027 +0000 UTC m=+1428.765045477" watchObservedRunningTime="2025-10-02 11:15:46.721554444 +0000 UTC m=+1428.775780894" Oct 02 11:15:47 crc kubenswrapper[4751]: I1002 11:15:47.565064 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af" path="/var/lib/kubelet/pods/4fc5a0c0-cbe6-4bae-9a82-5c9ed2f122af/volumes" Oct 02 11:15:47 crc kubenswrapper[4751]: I1002 11:15:47.565863 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6fac4cc-a2a9-420e-8179-0e835e1ddc9c" path="/var/lib/kubelet/pods/d6fac4cc-a2a9-420e-8179-0e835e1ddc9c/volumes" Oct 02 11:15:48 crc kubenswrapper[4751]: I1002 11:15:48.268458 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrsth" podUID="57a113ee-7eb3-4654-a0ad-48e5f69b6ed9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 02 11:15:49 crc kubenswrapper[4751]: I1002 11:15:49.691854 4751 generic.go:334] "Generic (PLEG): container finished" podID="f251a17b-99b0-4656-919b-1d03af5331a4" containerID="50836859af1ede94a8266a964e92ff2894489e3be518f679a8fe934fb90b06e4" exitCode=0 Oct 02 11:15:49 crc kubenswrapper[4751]: I1002 11:15:49.691990 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fbknj" event={"ID":"f251a17b-99b0-4656-919b-1d03af5331a4","Type":"ContainerDied","Data":"50836859af1ede94a8266a964e92ff2894489e3be518f679a8fe934fb90b06e4"} Oct 02 11:15:51 crc kubenswrapper[4751]: I1002 11:15:51.950687 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-44kgc" Oct 02 11:15:51 crc kubenswrapper[4751]: I1002 11:15:51.957420 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013421 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013476 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013508 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data\") pod \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013544 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013590 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data\") pod \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013620 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d98pt\" (UniqueName: \"kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt\") pod \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013658 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle\") pod \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\" (UID: \"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013688 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5xv2\" (UniqueName: \"kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013715 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.013794 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle\") pod \"f251a17b-99b0-4656-919b-1d03af5331a4\" (UID: \"f251a17b-99b0-4656-919b-1d03af5331a4\") " Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.018347 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts" (OuterVolumeSpecName: "scripts") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.018955 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.021493 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt" (OuterVolumeSpecName: "kube-api-access-d98pt") pod "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" (UID: "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c"). InnerVolumeSpecName "kube-api-access-d98pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.024406 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2" (OuterVolumeSpecName: "kube-api-access-n5xv2") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "kube-api-access-n5xv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.024519 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.027742 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" (UID: "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.047590 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.052119 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data" (OuterVolumeSpecName: "config-data") pod "f251a17b-99b0-4656-919b-1d03af5331a4" (UID: "f251a17b-99b0-4656-919b-1d03af5331a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.062571 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" (UID: "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.079028 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data" (OuterVolumeSpecName: "config-data") pod "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" (UID: "4d2a342d-289c-41e9-94f7-6ddbc10a6e1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115591 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5xv2\" (UniqueName: \"kubernetes.io/projected/f251a17b-99b0-4656-919b-1d03af5331a4-kube-api-access-n5xv2\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115627 4751 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115638 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115649 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115659 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115667 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115674 4751 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f251a17b-99b0-4656-919b-1d03af5331a4-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115682 4751 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115690 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d98pt\" (UniqueName: \"kubernetes.io/projected/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-kube-api-access-d98pt\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.115697 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.716803 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f50a9ca0-4300-4928-8d84-a174dd973a9f","Type":"ContainerStarted","Data":"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10"} Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.718906 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fbknj" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.718905 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fbknj" event={"ID":"f251a17b-99b0-4656-919b-1d03af5331a4","Type":"ContainerDied","Data":"08473cf3dbf7564d128107290e18898600fb8eaab40c9a11ac7132744d483774"} Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.719041 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08473cf3dbf7564d128107290e18898600fb8eaab40c9a11ac7132744d483774" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.722664 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-44kgc" event={"ID":"4d2a342d-289c-41e9-94f7-6ddbc10a6e1c","Type":"ContainerDied","Data":"0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0"} Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.722714 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6259c4c3fd617d982867a9514e2eef8918e6b79387c82b079d93ad1c50abf0" Oct 02 11:15:52 crc kubenswrapper[4751]: I1002 11:15:52.722783 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-44kgc" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.149501 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7b49c7b4b6-cxfxp"] Oct 02 11:15:53 crc kubenswrapper[4751]: E1002 11:15:53.150140 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" containerName="glance-db-sync" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.150152 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" containerName="glance-db-sync" Oct 02 11:15:53 crc kubenswrapper[4751]: E1002 11:15:53.150196 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f251a17b-99b0-4656-919b-1d03af5331a4" containerName="keystone-bootstrap" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.150203 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f251a17b-99b0-4656-919b-1d03af5331a4" containerName="keystone-bootstrap" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.150371 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" containerName="glance-db-sync" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.150390 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f251a17b-99b0-4656-919b-1d03af5331a4" containerName="keystone-bootstrap" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.150917 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.153652 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.153652 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.153867 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gk6gt" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.154430 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.155222 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.160088 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.178204 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b49c7b4b6-cxfxp"] Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233024 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-public-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233307 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-credential-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233422 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-scripts\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233518 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-fernet-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233633 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqxqf\" (UniqueName: \"kubernetes.io/projected/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-kube-api-access-qqxqf\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233852 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-combined-ca-bundle\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.233975 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-internal-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.234148 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-config-data\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.330200 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.330249 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.335952 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-credential-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336018 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-scripts\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336050 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-fernet-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336086 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqxqf\" (UniqueName: \"kubernetes.io/projected/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-kube-api-access-qqxqf\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336155 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-combined-ca-bundle\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336216 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-internal-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336279 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-config-data\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.336339 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-public-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.347129 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-credential-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.348732 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-internal-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.348887 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-public-tls-certs\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.349341 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-config-data\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.356273 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-combined-ca-bundle\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.360722 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-scripts\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.362710 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqxqf\" (UniqueName: \"kubernetes.io/projected/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-kube-api-access-qqxqf\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.375969 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f80ce64a-4124-4ce4-8ffe-66bdbaf3488b-fernet-keys\") pod \"keystone-7b49c7b4b6-cxfxp\" (UID: \"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b\") " pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.416549 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.418020 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.425304 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.426111 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.472002 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.485257 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.547774 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.547850 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.548016 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.548205 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5sff\" (UniqueName: \"kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.548314 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.548354 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.649968 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.650063 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.650139 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.650194 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.650270 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.650365 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5sff\" (UniqueName: \"kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.651624 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.651903 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.652540 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.653920 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.654356 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.677935 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5sff\" (UniqueName: \"kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff\") pod \"dnsmasq-dns-57c957c4ff-wnpfp\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.735433 4751 generic.go:334] "Generic (PLEG): container finished" podID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" containerID="119e0a5ecc875d6dce924f32efe9df3e575a54fe719c0861879bebe767dff91f" exitCode=0 Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.736261 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmv2n" event={"ID":"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479","Type":"ContainerDied","Data":"119e0a5ecc875d6dce924f32efe9df3e575a54fe719c0861879bebe767dff91f"} Oct 02 11:15:53 crc kubenswrapper[4751]: I1002 11:15:53.773481 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.138413 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b49c7b4b6-cxfxp"] Oct 02 11:15:54 crc kubenswrapper[4751]: W1002 11:15:54.145499 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf80ce64a_4124_4ce4_8ffe_66bdbaf3488b.slice/crio-297e54b21ad0ba1f9fb40d7d0497cc7d715ffac6818562570b04c481219d44dd WatchSource:0}: Error finding container 297e54b21ad0ba1f9fb40d7d0497cc7d715ffac6818562570b04c481219d44dd: Status 404 returned error can't find the container with id 297e54b21ad0ba1f9fb40d7d0497cc7d715ffac6818562570b04c481219d44dd Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.310599 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.321091 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.322607 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: W1002 11:15:54.325888 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc63ea0c_fa74_444f_8138_86d7f0145d7d.slice/crio-f049a2640a3ed546ed9344e913e654522e4f55defcc698f39b5b1bc90546c041 WatchSource:0}: Error finding container f049a2640a3ed546ed9344e913e654522e4f55defcc698f39b5b1bc90546c041: Status 404 returned error can't find the container with id f049a2640a3ed546ed9344e913e654522e4f55defcc698f39b5b1bc90546c041 Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.330543 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.330869 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xd7tx" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.330924 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.338159 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482444 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7n22\" (UniqueName: \"kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482532 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482588 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482789 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482865 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.482893 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.483044 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584597 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584747 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584794 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584826 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584845 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584910 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.584936 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7n22\" (UniqueName: \"kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.585204 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.585590 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.585652 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.592467 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.597872 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.604031 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.614188 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7n22\" (UniqueName: \"kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.620578 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.620844 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.623662 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.627268 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.638071 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.667243 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.747924 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b49c7b4b6-cxfxp" event={"ID":"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b","Type":"ContainerStarted","Data":"69b257ba65968477d724b5e592c34205f8674ccfab6c3d972ee48ff1c1155506"} Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.747983 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b49c7b4b6-cxfxp" event={"ID":"f80ce64a-4124-4ce4-8ffe-66bdbaf3488b","Type":"ContainerStarted","Data":"297e54b21ad0ba1f9fb40d7d0497cc7d715ffac6818562570b04c481219d44dd"} Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.748212 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.754451 4751 generic.go:334] "Generic (PLEG): container finished" podID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerID="339e663fa1f597a61c9de76234379ebf46b30773738b7501c277010f880adde3" exitCode=0 Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.754547 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" event={"ID":"bc63ea0c-fa74-444f-8138-86d7f0145d7d","Type":"ContainerDied","Data":"339e663fa1f597a61c9de76234379ebf46b30773738b7501c277010f880adde3"} Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.754583 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" event={"ID":"bc63ea0c-fa74-444f-8138-86d7f0145d7d","Type":"ContainerStarted","Data":"f049a2640a3ed546ed9344e913e654522e4f55defcc698f39b5b1bc90546c041"} Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.781518 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7b49c7b4b6-cxfxp" podStartSLOduration=1.781496941 podStartE2EDuration="1.781496941s" podCreationTimestamp="2025-10-02 11:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:54.772897761 +0000 UTC m=+1436.827124211" watchObservedRunningTime="2025-10-02 11:15:54.781496941 +0000 UTC m=+1436.835723411" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.789979 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790032 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790061 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790113 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790142 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xdwr\" (UniqueName: \"kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790212 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.790287 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.891852 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.892101 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.892145 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.893518 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.897129 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.904222 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xdwr\" (UniqueName: \"kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.904370 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.904545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.904717 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.909871 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.910777 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.910995 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.912051 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xdwr\" (UniqueName: \"kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.923807 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.941466 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:15:54 crc kubenswrapper[4751]: I1002 11:15:54.962334 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.308371 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.333943 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.415548 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle\") pod \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.415636 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data\") pod \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.415730 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvt4m\" (UniqueName: \"kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m\") pod \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\" (UID: \"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479\") " Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.421896 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m" (OuterVolumeSpecName: "kube-api-access-dvt4m") pod "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" (UID: "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479"). InnerVolumeSpecName "kube-api-access-dvt4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.431825 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" (UID: "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.460575 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" (UID: "d0a15cb7-4b3b-4b25-8b11-1b8f14b06479"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.517438 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.517475 4751 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.517484 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvt4m\" (UniqueName: \"kubernetes.io/projected/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479-kube-api-access-dvt4m\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.602475 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:15:55 crc kubenswrapper[4751]: W1002 11:15:55.619355 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf93f998d_110d_41f9_b2da_4731390ad62b.slice/crio-a23cb98178d0499748c505d279a58ee1071e12394bb493e824a53994371b45a6 WatchSource:0}: Error finding container a23cb98178d0499748c505d279a58ee1071e12394bb493e824a53994371b45a6: Status 404 returned error can't find the container with id a23cb98178d0499748c505d279a58ee1071e12394bb493e824a53994371b45a6 Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.769841 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerStarted","Data":"6cdb53a1a04b1eae789f05bb111479cd15d50280e514b6757983fe0f8fa8a5a8"} Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.775136 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerStarted","Data":"a23cb98178d0499748c505d279a58ee1071e12394bb493e824a53994371b45a6"} Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.780012 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" event={"ID":"bc63ea0c-fa74-444f-8138-86d7f0145d7d","Type":"ContainerStarted","Data":"b8ecd76bcb5114b2215ba9403e4af172cf6f1bae0e3bfe9d1a6304361584b042"} Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.780987 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.788481 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jmv2n" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.788539 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jmv2n" event={"ID":"d0a15cb7-4b3b-4b25-8b11-1b8f14b06479","Type":"ContainerDied","Data":"2635bf30974601800c5102c776315a12a9b5f6545f22dd4a332a23170fa0f0a4"} Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.788602 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2635bf30974601800c5102c776315a12a9b5f6545f22dd4a332a23170fa0f0a4" Oct 02 11:15:55 crc kubenswrapper[4751]: I1002 11:15:55.805881 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" podStartSLOduration=2.805865024 podStartE2EDuration="2.805865024s" podCreationTimestamp="2025-10-02 11:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:55.805606077 +0000 UTC m=+1437.859832537" watchObservedRunningTime="2025-10-02 11:15:55.805865024 +0000 UTC m=+1437.860091464" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.069931 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7c77b8fdb7-bhwjp"] Oct 02 11:15:56 crc kubenswrapper[4751]: E1002 11:15:56.073143 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" containerName="barbican-db-sync" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.082141 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" containerName="barbican-db-sync" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.082986 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" containerName="barbican-db-sync" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.127379 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.131183 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.134526 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4mw5f" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.134870 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.135035 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.135283 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c77b8fdb7-bhwjp"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.135379 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.147224 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.148144 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.223310 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241440 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-combined-ca-bundle\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241497 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data-custom\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241528 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-combined-ca-bundle\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241548 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8j6j\" (UniqueName: \"kubernetes.io/projected/cb83ae02-2a40-4786-a2aa-159f7c5a9141-kube-api-access-f8j6j\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241636 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kwgx\" (UniqueName: \"kubernetes.io/projected/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-kube-api-access-8kwgx\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241727 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241808 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb83ae02-2a40-4786-a2aa-159f7c5a9141-logs\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241834 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241855 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data-custom\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.241878 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-logs\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.247241 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.248703 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.260217 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343432 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb83ae02-2a40-4786-a2aa-159f7c5a9141-logs\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343502 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343524 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343548 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data-custom\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343575 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-logs\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343630 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-combined-ca-bundle\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343649 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data-custom\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343669 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-combined-ca-bundle\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343685 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8j6j\" (UniqueName: \"kubernetes.io/projected/cb83ae02-2a40-4786-a2aa-159f7c5a9141-kube-api-access-f8j6j\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343731 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343766 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kwgx\" (UniqueName: \"kubernetes.io/projected/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-kube-api-access-8kwgx\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343782 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343796 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb7rf\" (UniqueName: \"kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343819 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.343854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.344704 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb83ae02-2a40-4786-a2aa-159f7c5a9141-logs\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.350813 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-logs\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.355575 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.361243 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-combined-ca-bundle\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.361380 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.362780 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.369062 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data-custom\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.369411 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.373038 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-combined-ca-bundle\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.389324 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-config-data-custom\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.402692 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb83ae02-2a40-4786-a2aa-159f7c5a9141-config-data\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.403698 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kwgx\" (UniqueName: \"kubernetes.io/projected/c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38-kube-api-access-8kwgx\") pod \"barbican-keystone-listener-5d89ffdd4b-n5d8c\" (UID: \"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38\") " pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.418818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8j6j\" (UniqueName: \"kubernetes.io/projected/cb83ae02-2a40-4786-a2aa-159f7c5a9141-kube-api-access-f8j6j\") pod \"barbican-worker-7c77b8fdb7-bhwjp\" (UID: \"cb83ae02-2a40-4786-a2aa-159f7c5a9141\") " pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.437222 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448130 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448189 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb7rf\" (UniqueName: \"kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448218 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448257 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448313 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448339 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448369 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448397 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br7wl\" (UniqueName: \"kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448422 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.448466 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.449349 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.449899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.450851 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.451887 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.452863 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.479047 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb7rf\" (UniqueName: \"kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf\") pod \"dnsmasq-dns-6d66f584d7-29wnq\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.553109 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.553218 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br7wl\" (UniqueName: \"kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.553250 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.553366 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.553415 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.554076 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.557658 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.557855 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.558791 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.560302 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.565868 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.573893 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br7wl\" (UniqueName: \"kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl\") pod \"barbican-api-8699984c78-ls56z\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.587693 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.739637 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.913889 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.914880 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerStarted","Data":"98e8590bd90602f4c1b2842c20065f1838d4ad3ea77014f6296711fa0745aa9b"} Oct 02 11:15:56 crc kubenswrapper[4751]: I1002 11:15:56.933348 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerStarted","Data":"6c5210cea1eac5b6da773c41ea21d7d0baa8f3a64adfcc6bbb7720f9b9e8331e"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.022443 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.114928 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:15:57 crc kubenswrapper[4751]: W1002 11:15:57.115654 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fbe665b_0b68_4c7c_8e45_51685cee78cd.slice/crio-a6ca6ccffb381248560cb13a66245795a2625ceecdecfa13cb7d750c4e95ccc0 WatchSource:0}: Error finding container a6ca6ccffb381248560cb13a66245795a2625ceecdecfa13cb7d750c4e95ccc0: Status 404 returned error can't find the container with id a6ca6ccffb381248560cb13a66245795a2625ceecdecfa13cb7d750c4e95ccc0 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.153239 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c77b8fdb7-bhwjp"] Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.413529 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c"] Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.572065 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:15:57 crc kubenswrapper[4751]: W1002 11:15:57.579771 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda958d81b_34ed_444c_9717_28cde85f8bad.slice/crio-26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3 WatchSource:0}: Error finding container 26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3: Status 404 returned error can't find the container with id 26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.945808 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerStarted","Data":"26ff7860c125c5eed32ec953dff68a47517c8c4e3a58993780b074cdf1911572"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.945982 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-log" containerID="cri-o://98e8590bd90602f4c1b2842c20065f1838d4ad3ea77014f6296711fa0745aa9b" gracePeriod=30 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.946530 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-httpd" containerID="cri-o://26ff7860c125c5eed32ec953dff68a47517c8c4e3a58993780b074cdf1911572" gracePeriod=30 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.950404 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerStarted","Data":"625004d5cca52df557f82a0f68fbc2519c2bc19002a709f781ab72decd8ab14b"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.950499 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-log" containerID="cri-o://6c5210cea1eac5b6da773c41ea21d7d0baa8f3a64adfcc6bbb7720f9b9e8331e" gracePeriod=30 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.950574 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-httpd" containerID="cri-o://625004d5cca52df557f82a0f68fbc2519c2bc19002a709f781ab72decd8ab14b" gracePeriod=30 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.954601 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" event={"ID":"cb83ae02-2a40-4786-a2aa-159f7c5a9141","Type":"ContainerStarted","Data":"bc245222d5ba9d9a42756b5bab7ff7ecbfd16c6ca07b656e6efbe1437780c913"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.956626 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" event={"ID":"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38","Type":"ContainerStarted","Data":"a1e4264dfcae7700c08c3b070fe754c65b15c53b6b542152c11252672c246180"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.959070 4751 generic.go:334] "Generic (PLEG): container finished" podID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerID="077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251" exitCode=0 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.959146 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" event={"ID":"6fbe665b-0b68-4c7c-8e45-51685cee78cd","Type":"ContainerDied","Data":"077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.959229 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" event={"ID":"6fbe665b-0b68-4c7c-8e45-51685cee78cd","Type":"ContainerStarted","Data":"a6ca6ccffb381248560cb13a66245795a2625ceecdecfa13cb7d750c4e95ccc0"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.965460 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="dnsmasq-dns" containerID="cri-o://b8ecd76bcb5114b2215ba9403e4af172cf6f1bae0e3bfe9d1a6304361584b042" gracePeriod=10 Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.965542 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerStarted","Data":"4e937be64e941642a19a4590d6a8723cfc3aa2b7aefe489fa23a72b8dd421c05"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.965566 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerStarted","Data":"26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3"} Oct 02 11:15:57 crc kubenswrapper[4751]: I1002 11:15:57.980400 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.980378323 podStartE2EDuration="4.980378323s" podCreationTimestamp="2025-10-02 11:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:57.970240072 +0000 UTC m=+1440.024466542" watchObservedRunningTime="2025-10-02 11:15:57.980378323 +0000 UTC m=+1440.034604773" Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.008677 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.008651079 podStartE2EDuration="5.008651079s" podCreationTimestamp="2025-10-02 11:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:57.997947752 +0000 UTC m=+1440.052174212" watchObservedRunningTime="2025-10-02 11:15:58.008651079 +0000 UTC m=+1440.062877529" Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.993717 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerStarted","Data":"4da612369578da04394d2577ad56d3f64936b0b1e67eca23f4cf4d3e2eed7d51"} Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.994051 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.994069 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.998941 4751 generic.go:334] "Generic (PLEG): container finished" podID="4204d63c-be96-4651-8f3a-736101bf96ee" containerID="26ff7860c125c5eed32ec953dff68a47517c8c4e3a58993780b074cdf1911572" exitCode=0 Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.999009 4751 generic.go:334] "Generic (PLEG): container finished" podID="4204d63c-be96-4651-8f3a-736101bf96ee" containerID="98e8590bd90602f4c1b2842c20065f1838d4ad3ea77014f6296711fa0745aa9b" exitCode=143 Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.999076 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerDied","Data":"26ff7860c125c5eed32ec953dff68a47517c8c4e3a58993780b074cdf1911572"} Oct 02 11:15:58 crc kubenswrapper[4751]: I1002 11:15:58.999109 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerDied","Data":"98e8590bd90602f4c1b2842c20065f1838d4ad3ea77014f6296711fa0745aa9b"} Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.001991 4751 generic.go:334] "Generic (PLEG): container finished" podID="f93f998d-110d-41f9-b2da-4731390ad62b" containerID="625004d5cca52df557f82a0f68fbc2519c2bc19002a709f781ab72decd8ab14b" exitCode=0 Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.002022 4751 generic.go:334] "Generic (PLEG): container finished" podID="f93f998d-110d-41f9-b2da-4731390ad62b" containerID="6c5210cea1eac5b6da773c41ea21d7d0baa8f3a64adfcc6bbb7720f9b9e8331e" exitCode=143 Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.002068 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerDied","Data":"625004d5cca52df557f82a0f68fbc2519c2bc19002a709f781ab72decd8ab14b"} Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.002096 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerDied","Data":"6c5210cea1eac5b6da773c41ea21d7d0baa8f3a64adfcc6bbb7720f9b9e8331e"} Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.005571 4751 generic.go:334] "Generic (PLEG): container finished" podID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerID="b8ecd76bcb5114b2215ba9403e4af172cf6f1bae0e3bfe9d1a6304361584b042" exitCode=0 Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.005695 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" event={"ID":"bc63ea0c-fa74-444f-8138-86d7f0145d7d","Type":"ContainerDied","Data":"b8ecd76bcb5114b2215ba9403e4af172cf6f1bae0e3bfe9d1a6304361584b042"} Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.018843 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8699984c78-ls56z" podStartSLOduration=3.018814041 podStartE2EDuration="3.018814041s" podCreationTimestamp="2025-10-02 11:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:59.011948298 +0000 UTC m=+1441.066174768" watchObservedRunningTime="2025-10-02 11:15:59.018814041 +0000 UTC m=+1441.073040491" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.022207 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" event={"ID":"6fbe665b-0b68-4c7c-8e45-51685cee78cd","Type":"ContainerStarted","Data":"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315"} Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.022421 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.276013 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.284823 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.308920 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" podStartSLOduration=3.308898456 podStartE2EDuration="3.308898456s" podCreationTimestamp="2025-10-02 11:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:15:59.042354171 +0000 UTC m=+1441.096580631" watchObservedRunningTime="2025-10-02 11:15:59.308898456 +0000 UTC m=+1441.363124916" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325492 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325553 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325576 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325605 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325631 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325676 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.325780 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xdwr\" (UniqueName: \"kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr\") pod \"f93f998d-110d-41f9-b2da-4731390ad62b\" (UID: \"f93f998d-110d-41f9-b2da-4731390ad62b\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.326241 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs" (OuterVolumeSpecName: "logs") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.326891 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.336199 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr" (OuterVolumeSpecName: "kube-api-access-5xdwr") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "kube-api-access-5xdwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.345145 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts" (OuterVolumeSpecName: "scripts") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.358665 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.362083 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.417907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data" (OuterVolumeSpecName: "config-data") pod "f93f998d-110d-41f9-b2da-4731390ad62b" (UID: "f93f998d-110d-41f9-b2da-4731390ad62b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.427475 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.428059 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.428131 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.428298 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.428364 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.428548 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5sff\" (UniqueName: \"kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff\") pod \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\" (UID: \"bc63ea0c-fa74-444f-8138-86d7f0145d7d\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429266 4751 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429292 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429309 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93f998d-110d-41f9-b2da-4731390ad62b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429323 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429350 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429364 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93f998d-110d-41f9-b2da-4731390ad62b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.429376 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xdwr\" (UniqueName: \"kubernetes.io/projected/f93f998d-110d-41f9-b2da-4731390ad62b-kube-api-access-5xdwr\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.448237 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff" (OuterVolumeSpecName: "kube-api-access-k5sff") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "kube-api-access-k5sff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.457015 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.486266 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.487086 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.487567 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.504766 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config" (OuterVolumeSpecName: "config") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.513400 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc63ea0c-fa74-444f-8138-86d7f0145d7d" (UID: "bc63ea0c-fa74-444f-8138-86d7f0145d7d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530428 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530453 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530462 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530501 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530513 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530524 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc63ea0c-fa74-444f-8138-86d7f0145d7d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.530535 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5sff\" (UniqueName: \"kubernetes.io/projected/bc63ea0c-fa74-444f-8138-86d7f0145d7d-kube-api-access-k5sff\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.778166 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835186 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835268 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835376 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835459 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835507 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835617 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7n22\" (UniqueName: \"kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.835654 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs\") pod \"4204d63c-be96-4651-8f3a-736101bf96ee\" (UID: \"4204d63c-be96-4651-8f3a-736101bf96ee\") " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.836147 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.836334 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs" (OuterVolumeSpecName: "logs") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.842732 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22" (OuterVolumeSpecName: "kube-api-access-f7n22") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "kube-api-access-f7n22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.843729 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.843729 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts" (OuterVolumeSpecName: "scripts") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.876258 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.889471 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data" (OuterVolumeSpecName: "config-data") pod "4204d63c-be96-4651-8f3a-736101bf96ee" (UID: "4204d63c-be96-4651-8f3a-736101bf96ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938074 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938116 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7n22\" (UniqueName: \"kubernetes.io/projected/4204d63c-be96-4651-8f3a-736101bf96ee-kube-api-access-f7n22\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938131 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938143 4751 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4204d63c-be96-4651-8f3a-736101bf96ee-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938181 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938196 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.938210 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4204d63c-be96-4651-8f3a-736101bf96ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:59 crc kubenswrapper[4751]: I1002 11:15:59.963794 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.039444 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.071421 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.071721 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f93f998d-110d-41f9-b2da-4731390ad62b","Type":"ContainerDied","Data":"a23cb98178d0499748c505d279a58ee1071e12394bb493e824a53994371b45a6"} Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.071800 4751 scope.go:117] "RemoveContainer" containerID="625004d5cca52df557f82a0f68fbc2519c2bc19002a709f781ab72decd8ab14b" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.091858 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" event={"ID":"bc63ea0c-fa74-444f-8138-86d7f0145d7d","Type":"ContainerDied","Data":"f049a2640a3ed546ed9344e913e654522e4f55defcc698f39b5b1bc90546c041"} Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.092046 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wnpfp" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.104427 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.105065 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4204d63c-be96-4651-8f3a-736101bf96ee","Type":"ContainerDied","Data":"6cdb53a1a04b1eae789f05bb111479cd15d50280e514b6757983fe0f8fa8a5a8"} Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.125071 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.133942 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.155710 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.177267 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wnpfp"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186355 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.186864 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186889 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.186910 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186918 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.186945 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186952 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.186966 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="init" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186973 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="init" Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.186986 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.186993 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: E1002 11:16:00.187007 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="dnsmasq-dns" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187015 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="dnsmasq-dns" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187225 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187243 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" containerName="dnsmasq-dns" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187258 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187281 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" containerName="glance-log" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.187289 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" containerName="glance-httpd" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.188434 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.201003 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.201474 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.201675 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.201846 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xd7tx" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.202559 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.217776 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.227892 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.238697 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243087 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5brxm\" (UniqueName: \"kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243193 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243259 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243302 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243339 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243481 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243513 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.243541 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.244913 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.258995 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.258995 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.268534 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.303990 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5565ddb6f8-hdw9n"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.305976 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.310722 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.311201 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345046 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345087 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345109 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345127 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345417 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.345812 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.346758 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5565ddb6f8-hdw9n"] Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.347719 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-combined-ca-bundle\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.347753 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg7rs\" (UniqueName: \"kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.347854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data-custom\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.347918 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.347980 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5brxm\" (UniqueName: \"kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348003 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-internal-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348026 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348059 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348079 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4115615-cab2-4426-ba9f-6bb96b1fae57-logs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348142 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-public-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348183 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348250 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348269 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348289 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls9w8\" (UniqueName: \"kubernetes.io/projected/a4115615-cab2-4426-ba9f-6bb96b1fae57-kube-api-access-ls9w8\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348323 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348343 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348384 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.348440 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.350433 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.350917 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.354024 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.355130 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.355379 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.369026 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5brxm\" (UniqueName: \"kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.393937 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452414 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data-custom\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452470 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452503 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-internal-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452521 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4115615-cab2-4426-ba9f-6bb96b1fae57-logs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452571 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-public-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452587 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452606 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452620 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452639 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls9w8\" (UniqueName: \"kubernetes.io/projected/a4115615-cab2-4426-ba9f-6bb96b1fae57-kube-api-access-ls9w8\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452663 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452700 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452736 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452778 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-combined-ca-bundle\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.452795 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg7rs\" (UniqueName: \"kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.453057 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.457685 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.457864 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.460232 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4115615-cab2-4426-ba9f-6bb96b1fae57-logs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.469060 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data-custom\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.474510 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.481000 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-combined-ca-bundle\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.482113 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.495887 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-public-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.496410 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-internal-tls-certs\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.496537 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4115615-cab2-4426-ba9f-6bb96b1fae57-config-data\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.500235 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.507614 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls9w8\" (UniqueName: \"kubernetes.io/projected/a4115615-cab2-4426-ba9f-6bb96b1fae57-kube-api-access-ls9w8\") pod \"barbican-api-5565ddb6f8-hdw9n\" (UID: \"a4115615-cab2-4426-ba9f-6bb96b1fae57\") " pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.512509 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.527469 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg7rs\" (UniqueName: \"kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.527972 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.557004 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.585245 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:16:00 crc kubenswrapper[4751]: I1002 11:16:00.632645 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:01 crc kubenswrapper[4751]: I1002 11:16:01.568647 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4204d63c-be96-4651-8f3a-736101bf96ee" path="/var/lib/kubelet/pods/4204d63c-be96-4651-8f3a-736101bf96ee/volumes" Oct 02 11:16:01 crc kubenswrapper[4751]: I1002 11:16:01.570240 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc63ea0c-fa74-444f-8138-86d7f0145d7d" path="/var/lib/kubelet/pods/bc63ea0c-fa74-444f-8138-86d7f0145d7d/volumes" Oct 02 11:16:01 crc kubenswrapper[4751]: I1002 11:16:01.571082 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f93f998d-110d-41f9-b2da-4731390ad62b" path="/var/lib/kubelet/pods/f93f998d-110d-41f9-b2da-4731390ad62b/volumes" Oct 02 11:16:03 crc kubenswrapper[4751]: I1002 11:16:03.334349 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 11:16:03 crc kubenswrapper[4751]: I1002 11:16:03.427839 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6cf9f465b8-rf9bv" podUID="3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Oct 02 11:16:06 crc kubenswrapper[4751]: I1002 11:16:06.589415 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:16:06 crc kubenswrapper[4751]: I1002 11:16:06.678271 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:16:06 crc kubenswrapper[4751]: I1002 11:16:06.678822 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="dnsmasq-dns" containerID="cri-o://7fa9e8e4ed0c632999b45ce22cf369ceb740723c3d095862ba913072af850446" gracePeriod=10 Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.066694 4751 scope.go:117] "RemoveContainer" containerID="6c5210cea1eac5b6da773c41ea21d7d0baa8f3a64adfcc6bbb7720f9b9e8331e" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.171296 4751 generic.go:334] "Generic (PLEG): container finished" podID="148c43ee-5684-4168-8590-5128723f150c" containerID="7fa9e8e4ed0c632999b45ce22cf369ceb740723c3d095862ba913072af850446" exitCode=0 Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.171518 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" event={"ID":"148c43ee-5684-4168-8590-5128723f150c","Type":"ContainerDied","Data":"7fa9e8e4ed0c632999b45ce22cf369ceb740723c3d095862ba913072af850446"} Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.476647 4751 scope.go:117] "RemoveContainer" containerID="b8ecd76bcb5114b2215ba9403e4af172cf6f1bae0e3bfe9d1a6304361584b042" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.575338 4751 scope.go:117] "RemoveContainer" containerID="339e663fa1f597a61c9de76234379ebf46b30773738b7501c277010f880adde3" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.696832 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.716687 4751 scope.go:117] "RemoveContainer" containerID="26ff7860c125c5eed32ec953dff68a47517c8c4e3a58993780b074cdf1911572" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.818517 4751 scope.go:117] "RemoveContainer" containerID="98e8590bd90602f4c1b2842c20065f1838d4ad3ea77014f6296711fa0745aa9b" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820237 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820295 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820474 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820534 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820587 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.820706 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvjkk\" (UniqueName: \"kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk\") pod \"148c43ee-5684-4168-8590-5128723f150c\" (UID: \"148c43ee-5684-4168-8590-5128723f150c\") " Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.849123 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk" (OuterVolumeSpecName: "kube-api-access-jvjkk") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "kube-api-access-jvjkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:07 crc kubenswrapper[4751]: I1002 11:16:07.922668 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvjkk\" (UniqueName: \"kubernetes.io/projected/148c43ee-5684-4168-8590-5128723f150c-kube-api-access-jvjkk\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.120723 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.144354 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config" (OuterVolumeSpecName: "config") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.170730 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.170945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.178482 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.228759 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.228791 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.228805 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.228816 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.238623 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" event={"ID":"cb83ae02-2a40-4786-a2aa-159f7c5a9141","Type":"ContainerStarted","Data":"a0b96bd15404c1e639870c89f0e2a53561f17166ee1f8a9e273b63936846cbfb"} Oct 02 11:16:08 crc kubenswrapper[4751]: E1002 11:16:08.246761 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"ceilometer-notification-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.251957 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" event={"ID":"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38","Type":"ContainerStarted","Data":"5f084af2ce7fdc4778a8741410ae9ed269ec5fba55b8061db618291b99eef12d"} Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.253127 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "148c43ee-5684-4168-8590-5128723f150c" (UID: "148c43ee-5684-4168-8590-5128723f150c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.259066 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerStarted","Data":"c987f12462ec5e8259dee7a0b8437776673588280bb6ba3ffa1a9bbf4f7cac85"} Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.260743 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.269667 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5565ddb6f8-hdw9n"] Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.283887 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" event={"ID":"148c43ee-5684-4168-8590-5128723f150c","Type":"ContainerDied","Data":"8e87cf0c5f7994844a3c050f7eedd34b53157e27ff44e8a6fcea2f6a1ed73b22"} Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.283998 4751 scope.go:117] "RemoveContainer" containerID="7fa9e8e4ed0c632999b45ce22cf369ceb740723c3d095862ba913072af850446" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.284073 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-q25mx" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.305709 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-247rg" event={"ID":"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471","Type":"ContainerStarted","Data":"8703551d47e0a33f9113189ab52914c5416d36420ba174c5d2cada59c8f8dea1"} Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.322238 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="sg-core" containerID="cri-o://e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" gracePeriod=30 Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.332133 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.332286 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="proxy-httpd" containerID="cri-o://0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" gracePeriod=30 Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.337303 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/148c43ee-5684-4168-8590-5128723f150c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:08 crc kubenswrapper[4751]: W1002 11:16:08.343773 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4115615_cab2_4426_ba9f_6bb96b1fae57.slice/crio-dfcb755e10849ef524dc38c6f12c042b34014ebbeba2a27c93313c8e1302d54c WatchSource:0}: Error finding container dfcb755e10849ef524dc38c6f12c042b34014ebbeba2a27c93313c8e1302d54c: Status 404 returned error can't find the container with id dfcb755e10849ef524dc38c6f12c042b34014ebbeba2a27c93313c8e1302d54c Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.386863 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-247rg" podStartSLOduration=5.492534917 podStartE2EDuration="1m14.386840216s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="2025-10-02 11:14:55.827155535 +0000 UTC m=+1377.881381985" lastFinishedPulling="2025-10-02 11:16:04.721460834 +0000 UTC m=+1446.775687284" observedRunningTime="2025-10-02 11:16:08.332870413 +0000 UTC m=+1450.387096863" watchObservedRunningTime="2025-10-02 11:16:08.386840216 +0000 UTC m=+1450.441066676" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.478311 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.482710 4751 scope.go:117] "RemoveContainer" containerID="b7662a2245f781b0a60ef019acbee9cf0f655a428ef4cb41935c353d104d7224" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.492733 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-q25mx"] Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.932512 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:08 crc kubenswrapper[4751]: I1002 11:16:08.997267 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069321 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069385 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069480 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069528 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069556 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069638 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069728 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jdbf\" (UniqueName: \"kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf\") pod \"f50a9ca0-4300-4928-8d84-a174dd973a9f\" (UID: \"f50a9ca0-4300-4928-8d84-a174dd973a9f\") " Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069837 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.069983 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.070284 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.070311 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f50a9ca0-4300-4928-8d84-a174dd973a9f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.074591 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf" (OuterVolumeSpecName: "kube-api-access-9jdbf") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "kube-api-access-9jdbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.079103 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts" (OuterVolumeSpecName: "scripts") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.121688 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.140328 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.153344 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data" (OuterVolumeSpecName: "config-data") pod "f50a9ca0-4300-4928-8d84-a174dd973a9f" (UID: "f50a9ca0-4300-4928-8d84-a174dd973a9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.172903 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.172940 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jdbf\" (UniqueName: \"kubernetes.io/projected/f50a9ca0-4300-4928-8d84-a174dd973a9f-kube-api-access-9jdbf\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.172952 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.172963 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.172972 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f50a9ca0-4300-4928-8d84-a174dd973a9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.353683 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28csr" event={"ID":"faae59de-da62-46ae-b58e-acb43ada5fec","Type":"ContainerStarted","Data":"2c701b07095f80e46dcb30b2fb96048e33d67f797796f9b03d7ce01770fc0a5b"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376632 4751 generic.go:334] "Generic (PLEG): container finished" podID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerID="0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" exitCode=0 Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376666 4751 generic.go:334] "Generic (PLEG): container finished" podID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerID="e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" exitCode=2 Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376739 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376792 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f50a9ca0-4300-4928-8d84-a174dd973a9f","Type":"ContainerDied","Data":"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376823 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f50a9ca0-4300-4928-8d84-a174dd973a9f","Type":"ContainerDied","Data":"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376836 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f50a9ca0-4300-4928-8d84-a174dd973a9f","Type":"ContainerDied","Data":"43e66679553c4834b21b4d579d8d5febd25cbabe442cef0e07e621a5e8e98905"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.376850 4751 scope.go:117] "RemoveContainer" containerID="0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.381680 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5565ddb6f8-hdw9n" event={"ID":"a4115615-cab2-4426-ba9f-6bb96b1fae57","Type":"ContainerStarted","Data":"b32b4d79b0609d92282e6b67cbd904867d27844d73d8fdda0c06f69a74c5f64a"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.381707 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5565ddb6f8-hdw9n" event={"ID":"a4115615-cab2-4426-ba9f-6bb96b1fae57","Type":"ContainerStarted","Data":"dfcb755e10849ef524dc38c6f12c042b34014ebbeba2a27c93313c8e1302d54c"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.381865 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-28csr" podStartSLOduration=3.775808467 podStartE2EDuration="1m15.381856444s" podCreationTimestamp="2025-10-02 11:14:54 +0000 UTC" firstStartedPulling="2025-10-02 11:14:55.42479875 +0000 UTC m=+1377.479025200" lastFinishedPulling="2025-10-02 11:16:07.030846727 +0000 UTC m=+1449.085073177" observedRunningTime="2025-10-02 11:16:09.376631464 +0000 UTC m=+1451.430857934" watchObservedRunningTime="2025-10-02 11:16:09.381856444 +0000 UTC m=+1451.436082894" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.383012 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" event={"ID":"cb83ae02-2a40-4786-a2aa-159f7c5a9141","Type":"ContainerStarted","Data":"516a7fcc41d11e375c5918add1f3bdbd86cc3849d40108673895751d5b6a3dbe"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.385599 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerStarted","Data":"1be9726ec684bf1608b543fe0251e00b8de0c7ebfac8cc33e2442aadd9332ed2"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.399589 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" event={"ID":"c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38","Type":"ContainerStarted","Data":"e6537e7db3e5b40cf9c6caf3353b192674bdde03f64881d049c0681c40eed0b3"} Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.437030 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7c77b8fdb7-bhwjp" podStartSLOduration=4.405225713 podStartE2EDuration="13.437007638s" podCreationTimestamp="2025-10-02 11:15:56 +0000 UTC" firstStartedPulling="2025-10-02 11:15:57.167588545 +0000 UTC m=+1439.221814985" lastFinishedPulling="2025-10-02 11:16:06.19937046 +0000 UTC m=+1448.253596910" observedRunningTime="2025-10-02 11:16:09.433155805 +0000 UTC m=+1451.487382265" watchObservedRunningTime="2025-10-02 11:16:09.437007638 +0000 UTC m=+1451.491234088" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.438521 4751 scope.go:117] "RemoveContainer" containerID="e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.492136 4751 scope.go:117] "RemoveContainer" containerID="0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.492644 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f\": container with ID starting with 0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f not found: ID does not exist" containerID="0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.492705 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f"} err="failed to get container status \"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f\": rpc error: code = NotFound desc = could not find container \"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f\": container with ID starting with 0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f not found: ID does not exist" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.492741 4751 scope.go:117] "RemoveContainer" containerID="e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.493051 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10\": container with ID starting with e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10 not found: ID does not exist" containerID="e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.493086 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10"} err="failed to get container status \"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10\": rpc error: code = NotFound desc = could not find container \"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10\": container with ID starting with e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10 not found: ID does not exist" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.493107 4751 scope.go:117] "RemoveContainer" containerID="0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.493377 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f"} err="failed to get container status \"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f\": rpc error: code = NotFound desc = could not find container \"0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f\": container with ID starting with 0354b59aef70b4fd409cb799cfdb261d47e3d65e2ef1eae524f7712429beea3f not found: ID does not exist" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.493401 4751 scope.go:117] "RemoveContainer" containerID="e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.493595 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10"} err="failed to get container status \"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10\": rpc error: code = NotFound desc = could not find container \"e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10\": container with ID starting with e01d733788390602cdff8b82897da2720f85cb774ba1f0eed1eb927d27d76d10 not found: ID does not exist" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.521237 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5d89ffdd4b-n5d8c" podStartSLOduration=3.951778861 podStartE2EDuration="13.521212039s" podCreationTimestamp="2025-10-02 11:15:56 +0000 UTC" firstStartedPulling="2025-10-02 11:15:57.498011288 +0000 UTC m=+1439.552237728" lastFinishedPulling="2025-10-02 11:16:07.067444456 +0000 UTC m=+1449.121670906" observedRunningTime="2025-10-02 11:16:09.460141656 +0000 UTC m=+1451.514368126" watchObservedRunningTime="2025-10-02 11:16:09.521212039 +0000 UTC m=+1451.575438489" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.537009 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.551231 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.584553 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="148c43ee-5684-4168-8590-5128723f150c" path="/var/lib/kubelet/pods/148c43ee-5684-4168-8590-5128723f150c/volumes" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.586545 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" path="/var/lib/kubelet/pods/f50a9ca0-4300-4928-8d84-a174dd973a9f/volumes" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.596602 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.597626 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="init" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.597761 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="init" Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.597891 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="sg-core" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.597963 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="sg-core" Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.598036 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="dnsmasq-dns" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.598110 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="dnsmasq-dns" Oct 02 11:16:09 crc kubenswrapper[4751]: E1002 11:16:09.599383 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="proxy-httpd" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.599540 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="proxy-httpd" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.599935 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="proxy-httpd" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.600687 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="148c43ee-5684-4168-8590-5128723f150c" containerName="dnsmasq-dns" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.600799 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f50a9ca0-4300-4928-8d84-a174dd973a9f" containerName="sg-core" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.608521 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.611431 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.614682 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.615568 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709519 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709591 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709691 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709736 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709898 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77sbx\" (UniqueName: \"kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709931 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.709969 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.811813 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812143 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812230 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77sbx\" (UniqueName: \"kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812257 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812280 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812375 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.812896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.815553 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.821077 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.822094 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.831792 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.836853 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.842152 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77sbx\" (UniqueName: \"kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx\") pod \"ceilometer-0\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " pod="openstack/ceilometer-0" Oct 02 11:16:09 crc kubenswrapper[4751]: I1002 11:16:09.929525 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.274340 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.408965 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerStarted","Data":"5cb1cdbfff03f8989608a29c15b35b95b8e7743d1f23ff10e9d5835d88718b2d"} Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.421487 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5565ddb6f8-hdw9n" event={"ID":"a4115615-cab2-4426-ba9f-6bb96b1fae57","Type":"ContainerStarted","Data":"f846ba54a4fa788e7a0eb1be3c36181830bc46cf23b43edf3270eae3c031d4ce"} Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.421678 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.421742 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.427485 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerStarted","Data":"7ed61b6e52b31cf20bff3c00c44e3f575bd256d061571e9f7d5e9ee34f23d14e"} Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.467357 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5565ddb6f8-hdw9n" podStartSLOduration=10.46733062 podStartE2EDuration="10.46733062s" podCreationTimestamp="2025-10-02 11:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:10.445432335 +0000 UTC m=+1452.499658805" watchObservedRunningTime="2025-10-02 11:16:10.46733062 +0000 UTC m=+1452.521557070" Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.693967 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:10 crc kubenswrapper[4751]: I1002 11:16:10.694724 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:16:11 crc kubenswrapper[4751]: I1002 11:16:11.446622 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerStarted","Data":"3756ebb7f16c937287a2186b0644bef120fff85f837fa23d716a59d3215738dd"} Oct 02 11:16:11 crc kubenswrapper[4751]: I1002 11:16:11.449907 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerStarted","Data":"11ea07032f163f0087cf6b0727e96c2757f81dd4a7ff8eb23db199c0c6b67ba7"} Oct 02 11:16:11 crc kubenswrapper[4751]: I1002 11:16:11.457331 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerStarted","Data":"4f7ff3de9998d2d398821bd2603fc30726046078935698f5fcc731974d93ecc2"} Oct 02 11:16:11 crc kubenswrapper[4751]: I1002 11:16:11.492355 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.492330081 podStartE2EDuration="11.492330081s" podCreationTimestamp="2025-10-02 11:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:11.477259048 +0000 UTC m=+1453.531485518" watchObservedRunningTime="2025-10-02 11:16:11.492330081 +0000 UTC m=+1453.546556551" Oct 02 11:16:11 crc kubenswrapper[4751]: I1002 11:16:11.523833 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.523808742 podStartE2EDuration="11.523808742s" podCreationTimestamp="2025-10-02 11:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:11.511441312 +0000 UTC m=+1453.565667782" watchObservedRunningTime="2025-10-02 11:16:11.523808742 +0000 UTC m=+1453.578035212" Oct 02 11:16:12 crc kubenswrapper[4751]: I1002 11:16:12.473009 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerStarted","Data":"2d8c61f970fb9a0daa43635f8d052ed1bb137f09fa033437f8fcafdb62b5fd95"} Oct 02 11:16:13 crc kubenswrapper[4751]: I1002 11:16:13.484606 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerStarted","Data":"e6767c6ef027c584acc5f0ac64d9ee81503b1d109b149ec70b48530d228c295b"} Oct 02 11:16:14 crc kubenswrapper[4751]: I1002 11:16:14.496269 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerStarted","Data":"8e8046a4f3c6192b5b6d2afbaba742a19e3e99048ab5f79ef748e1d34fc002d3"} Oct 02 11:16:15 crc kubenswrapper[4751]: I1002 11:16:15.133595 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:16:15 crc kubenswrapper[4751]: I1002 11:16:15.486178 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:16:16 crc kubenswrapper[4751]: I1002 11:16:16.515489 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerStarted","Data":"516af6a81663d2aac64b770664cf4d914871763212056fc5759773d959c5c60b"} Oct 02 11:16:16 crc kubenswrapper[4751]: I1002 11:16:16.515941 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.072371 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.097043 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.371139913 podStartE2EDuration="8.097021814s" podCreationTimestamp="2025-10-02 11:16:09 +0000 UTC" firstStartedPulling="2025-10-02 11:16:10.694492253 +0000 UTC m=+1452.748718703" lastFinishedPulling="2025-10-02 11:16:15.420374154 +0000 UTC m=+1457.474600604" observedRunningTime="2025-10-02 11:16:16.544449762 +0000 UTC m=+1458.598676242" watchObservedRunningTime="2025-10-02 11:16:17.097021814 +0000 UTC m=+1459.151248274" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.239072 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.444399 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5565ddb6f8-hdw9n" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.572223 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.572468 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8699984c78-ls56z" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api-log" containerID="cri-o://4e937be64e941642a19a4590d6a8723cfc3aa2b7aefe489fa23a72b8dd421c05" gracePeriod=30 Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.572827 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8699984c78-ls56z" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api" containerID="cri-o://4da612369578da04394d2577ad56d3f64936b0b1e67eca23f4cf4d3e2eed7d51" gracePeriod=30 Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.779033 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6cf9f465b8-rf9bv" Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.844415 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.844831 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" containerID="cri-o://d56322dba3fe24647b339aa90dbabe8be5b32654c39498af79c501dcd44d87ac" gracePeriod=30 Oct 02 11:16:17 crc kubenswrapper[4751]: I1002 11:16:17.844686 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon-log" containerID="cri-o://f351c5cafadb1cd76de3df169a58c548a4b6d6e4f9c992ba4f53d1b041ba6cdf" gracePeriod=30 Oct 02 11:16:18 crc kubenswrapper[4751]: I1002 11:16:18.555485 4751 generic.go:334] "Generic (PLEG): container finished" podID="a958d81b-34ed-444c-9717-28cde85f8bad" containerID="4e937be64e941642a19a4590d6a8723cfc3aa2b7aefe489fa23a72b8dd421c05" exitCode=143 Oct 02 11:16:18 crc kubenswrapper[4751]: I1002 11:16:18.555586 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerDied","Data":"4e937be64e941642a19a4590d6a8723cfc3aa2b7aefe489fa23a72b8dd421c05"} Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.528567 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.528916 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.564903 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.571755 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.577927 4751 generic.go:334] "Generic (PLEG): container finished" podID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" containerID="8703551d47e0a33f9113189ab52914c5416d36420ba174c5d2cada59c8f8dea1" exitCode=0 Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.578002 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-247rg" event={"ID":"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471","Type":"ContainerDied","Data":"8703551d47e0a33f9113189ab52914c5416d36420ba174c5d2cada59c8f8dea1"} Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.578440 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.578532 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.585567 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.585619 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.628682 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 11:16:20 crc kubenswrapper[4751]: I1002 11:16:20.664191 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.596800 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerDied","Data":"4da612369578da04394d2577ad56d3f64936b0b1e67eca23f4cf4d3e2eed7d51"} Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.596796 4751 generic.go:334] "Generic (PLEG): container finished" podID="a958d81b-34ed-444c-9717-28cde85f8bad" containerID="4da612369578da04394d2577ad56d3f64936b0b1e67eca23f4cf4d3e2eed7d51" exitCode=0 Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.597129 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8699984c78-ls56z" event={"ID":"a958d81b-34ed-444c-9717-28cde85f8bad","Type":"ContainerDied","Data":"26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3"} Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.597141 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26991312b9c75929d028ffd119711ec41cba0b012161b6e116ac043fa71fd7f3" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.601080 4751 generic.go:334] "Generic (PLEG): container finished" podID="61916e69-5770-4e34-acae-3b8a551df701" containerID="d56322dba3fe24647b339aa90dbabe8be5b32654c39498af79c501dcd44d87ac" exitCode=0 Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.601419 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerDied","Data":"d56322dba3fe24647b339aa90dbabe8be5b32654c39498af79c501dcd44d87ac"} Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.604446 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.604493 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.635535 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.664827 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle\") pod \"a958d81b-34ed-444c-9717-28cde85f8bad\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.665243 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs\") pod \"a958d81b-34ed-444c-9717-28cde85f8bad\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.665309 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br7wl\" (UniqueName: \"kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl\") pod \"a958d81b-34ed-444c-9717-28cde85f8bad\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.665349 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom\") pod \"a958d81b-34ed-444c-9717-28cde85f8bad\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.665404 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data\") pod \"a958d81b-34ed-444c-9717-28cde85f8bad\" (UID: \"a958d81b-34ed-444c-9717-28cde85f8bad\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.670649 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a958d81b-34ed-444c-9717-28cde85f8bad" (UID: "a958d81b-34ed-444c-9717-28cde85f8bad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.666015 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs" (OuterVolumeSpecName: "logs") pod "a958d81b-34ed-444c-9717-28cde85f8bad" (UID: "a958d81b-34ed-444c-9717-28cde85f8bad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.685377 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl" (OuterVolumeSpecName: "kube-api-access-br7wl") pod "a958d81b-34ed-444c-9717-28cde85f8bad" (UID: "a958d81b-34ed-444c-9717-28cde85f8bad"). InnerVolumeSpecName "kube-api-access-br7wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.704458 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a958d81b-34ed-444c-9717-28cde85f8bad" (UID: "a958d81b-34ed-444c-9717-28cde85f8bad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.749102 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data" (OuterVolumeSpecName: "config-data") pod "a958d81b-34ed-444c-9717-28cde85f8bad" (UID: "a958d81b-34ed-444c-9717-28cde85f8bad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.768037 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br7wl\" (UniqueName: \"kubernetes.io/projected/a958d81b-34ed-444c-9717-28cde85f8bad-kube-api-access-br7wl\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.768071 4751 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.768080 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.768089 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a958d81b-34ed-444c-9717-28cde85f8bad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.768100 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a958d81b-34ed-444c-9717-28cde85f8bad-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.950621 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-247rg" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.981390 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts\") pod \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.981447 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data\") pod \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.981639 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle\") pod \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.981670 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b98l4\" (UniqueName: \"kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4\") pod \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.981817 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs\") pod \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\" (UID: \"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471\") " Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.982536 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs" (OuterVolumeSpecName: "logs") pod "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" (UID: "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.987157 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts" (OuterVolumeSpecName: "scripts") pod "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" (UID: "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:21 crc kubenswrapper[4751]: I1002 11:16:21.987334 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4" (OuterVolumeSpecName: "kube-api-access-b98l4") pod "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" (UID: "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471"). InnerVolumeSpecName "kube-api-access-b98l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.014463 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data" (OuterVolumeSpecName: "config-data") pod "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" (UID: "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.019557 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" (UID: "c9bb5b1a-3008-4ca4-8c88-dac1b5b22471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:22 crc kubenswrapper[4751]: E1002 11:16:22.054482 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6080e003_0583_428a_b263_3badfde0fd37.slice/crio-conmon-a728fae2f1a380c15e5618cdda55c64f9b462f8bb159e26d29d607b390c303a3.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.086575 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.086609 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b98l4\" (UniqueName: \"kubernetes.io/projected/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-kube-api-access-b98l4\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.086621 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.086630 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.086638 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.614496 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-247rg" event={"ID":"c9bb5b1a-3008-4ca4-8c88-dac1b5b22471","Type":"ContainerDied","Data":"a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d"} Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.614634 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7090b090104d0bf525c8d5f576e285de659bd4605ee44f71f7c72d23400900d" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.614840 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-247rg" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.618808 4751 generic.go:334] "Generic (PLEG): container finished" podID="6080e003-0583-428a-b263-3badfde0fd37" containerID="a728fae2f1a380c15e5618cdda55c64f9b462f8bb159e26d29d607b390c303a3" exitCode=0 Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.618920 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m2fxt" event={"ID":"6080e003-0583-428a-b263-3badfde0fd37","Type":"ContainerDied","Data":"a728fae2f1a380c15e5618cdda55c64f9b462f8bb159e26d29d607b390c303a3"} Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.619648 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8699984c78-ls56z" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.658687 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.664431 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8699984c78-ls56z"] Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.753684 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6d856cc8dd-n7jjw"] Oct 02 11:16:22 crc kubenswrapper[4751]: E1002 11:16:22.754121 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" containerName="placement-db-sync" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754148 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" containerName="placement-db-sync" Oct 02 11:16:22 crc kubenswrapper[4751]: E1002 11:16:22.754167 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754191 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api" Oct 02 11:16:22 crc kubenswrapper[4751]: E1002 11:16:22.754211 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api-log" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754218 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api-log" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754409 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" containerName="placement-db-sync" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754429 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.754466 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" containerName="barbican-api-log" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.755854 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.757803 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.758998 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.759187 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.760249 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zpwmv" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.760434 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.766880 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d856cc8dd-n7jjw"] Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.768791 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.768927 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.771187 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.912192 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25b5\" (UniqueName: \"kubernetes.io/projected/c9200866-1ba8-48d2-b271-9b55482ea983-kube-api-access-l25b5\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.912252 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9200866-1ba8-48d2-b271-9b55482ea983-logs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.912324 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-scripts\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.912623 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-public-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.912714 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-internal-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.913081 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-config-data\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:22 crc kubenswrapper[4751]: I1002 11:16:22.913354 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-combined-ca-bundle\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.017265 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9200866-1ba8-48d2-b271-9b55482ea983-logs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.017699 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-scripts\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.017643 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9200866-1ba8-48d2-b271-9b55482ea983-logs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.017794 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-public-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.018338 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-internal-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.018397 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-config-data\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.018421 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-combined-ca-bundle\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.018505 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25b5\" (UniqueName: \"kubernetes.io/projected/c9200866-1ba8-48d2-b271-9b55482ea983-kube-api-access-l25b5\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.022052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-combined-ca-bundle\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.022065 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-internal-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.023419 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-public-tls-certs\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.026831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-scripts\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.033857 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9200866-1ba8-48d2-b271-9b55482ea983-config-data\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.041153 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25b5\" (UniqueName: \"kubernetes.io/projected/c9200866-1ba8-48d2-b271-9b55482ea983-kube-api-access-l25b5\") pod \"placement-6d856cc8dd-n7jjw\" (UID: \"c9200866-1ba8-48d2-b271-9b55482ea983\") " pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.081313 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.331343 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.561505 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a958d81b-34ed-444c-9717-28cde85f8bad" path="/var/lib/kubelet/pods/a958d81b-34ed-444c-9717-28cde85f8bad/volumes" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.607916 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d856cc8dd-n7jjw"] Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.634751 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d856cc8dd-n7jjw" event={"ID":"c9200866-1ba8-48d2-b271-9b55482ea983","Type":"ContainerStarted","Data":"8fa4de14d8474cf42cd029fc86741e18f11647806760acf6c9a8e4a6e133dbd4"} Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.634843 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.634856 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.824818 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 11:16:23 crc kubenswrapper[4751]: I1002 11:16:23.830876 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.039557 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.138569 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle\") pod \"6080e003-0583-428a-b263-3badfde0fd37\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.138670 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config\") pod \"6080e003-0583-428a-b263-3badfde0fd37\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.138754 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zxp9\" (UniqueName: \"kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9\") pod \"6080e003-0583-428a-b263-3badfde0fd37\" (UID: \"6080e003-0583-428a-b263-3badfde0fd37\") " Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.143503 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9" (OuterVolumeSpecName: "kube-api-access-5zxp9") pod "6080e003-0583-428a-b263-3badfde0fd37" (UID: "6080e003-0583-428a-b263-3badfde0fd37"). InnerVolumeSpecName "kube-api-access-5zxp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.177324 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6080e003-0583-428a-b263-3badfde0fd37" (UID: "6080e003-0583-428a-b263-3badfde0fd37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.183623 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config" (OuterVolumeSpecName: "config") pod "6080e003-0583-428a-b263-3badfde0fd37" (UID: "6080e003-0583-428a-b263-3badfde0fd37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.240524 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.240558 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6080e003-0583-428a-b263-3badfde0fd37-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.240579 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zxp9\" (UniqueName: \"kubernetes.io/projected/6080e003-0583-428a-b263-3badfde0fd37-kube-api-access-5zxp9\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.673566 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d856cc8dd-n7jjw" event={"ID":"c9200866-1ba8-48d2-b271-9b55482ea983","Type":"ContainerStarted","Data":"a5b9fd5e28a0bdd283fd0a746eee95f1a82ee3da2b851186b3bdbefe047b506e"} Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.673931 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d856cc8dd-n7jjw" event={"ID":"c9200866-1ba8-48d2-b271-9b55482ea983","Type":"ContainerStarted","Data":"187222e72c4617f6576fb32cf961614db2ff61d5e35cbc73d880ce7164c11b34"} Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.675150 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.675200 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.684591 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m2fxt" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.688781 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m2fxt" event={"ID":"6080e003-0583-428a-b263-3badfde0fd37","Type":"ContainerDied","Data":"cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d"} Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.688836 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd4851fc7020ff3d6f2d72f4b4ce33a631cbecd0cc615c96e6eb79095edffa5d" Oct 02 11:16:24 crc kubenswrapper[4751]: I1002 11:16:24.702760 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6d856cc8dd-n7jjw" podStartSLOduration=2.702741998 podStartE2EDuration="2.702741998s" podCreationTimestamp="2025-10-02 11:16:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:24.696654005 +0000 UTC m=+1466.750880465" watchObservedRunningTime="2025-10-02 11:16:24.702741998 +0000 UTC m=+1466.756968448" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.032582 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:25 crc kubenswrapper[4751]: E1002 11:16:25.033000 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6080e003-0583-428a-b263-3badfde0fd37" containerName="neutron-db-sync" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.033013 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6080e003-0583-428a-b263-3badfde0fd37" containerName="neutron-db-sync" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.033277 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6080e003-0583-428a-b263-3badfde0fd37" containerName="neutron-db-sync" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.034216 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.073636 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.133721 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.135228 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.137822 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-stfgf" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.138005 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.144256 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.144453 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.164921 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170528 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170591 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170640 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170666 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170693 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zth8g\" (UniqueName: \"kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.170726 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272039 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272154 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272258 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272293 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272329 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272553 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.272612 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.273494 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.273861 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274129 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274220 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274301 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274345 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zth8g\" (UniqueName: \"kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274382 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.274430 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9xrj\" (UniqueName: \"kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.275052 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.299131 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zth8g\" (UniqueName: \"kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g\") pod \"dnsmasq-dns-688c87cc99-ntlpt\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.369997 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.376002 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.376076 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.376158 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.376198 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9xrj\" (UniqueName: \"kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.376272 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.386704 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.387300 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.388109 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.394894 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.406693 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9xrj\" (UniqueName: \"kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj\") pod \"neutron-5d9c6c568d-bsnqp\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.460553 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:25 crc kubenswrapper[4751]: I1002 11:16:25.849639 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7b49c7b4b6-cxfxp" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.092998 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:26 crc kubenswrapper[4751]: W1002 11:16:26.096666 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod361e9060_8522_477e_b827_e2113097fa95.slice/crio-732f0558d91e28c1b6ff61a98720f98a1a8705d4ffa3e17d00e9e833794324e0 WatchSource:0}: Error finding container 732f0558d91e28c1b6ff61a98720f98a1a8705d4ffa3e17d00e9e833794324e0: Status 404 returned error can't find the container with id 732f0558d91e28c1b6ff61a98720f98a1a8705d4ffa3e17d00e9e833794324e0 Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.116420 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.182909 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.186394 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.189715 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.190682 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-lhq4x" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.192746 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.200483 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.303194 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.303313 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.303358 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xngm\" (UniqueName: \"kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.303550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.406377 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.406489 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.406635 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.406706 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xngm\" (UniqueName: \"kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.407804 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.410891 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.412791 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.424833 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xngm\" (UniqueName: \"kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm\") pod \"openstackclient\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.492779 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.493908 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.503446 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.530645 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.531995 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.546771 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.611637 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.611746 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.611816 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config-secret\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.611846 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz8qn\" (UniqueName: \"kubernetes.io/projected/4002bd9a-1957-499c-b077-3f1eee6b239e-kube-api-access-vz8qn\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: E1002 11:16:26.682529 4751 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 02 11:16:26 crc kubenswrapper[4751]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_17591eff-a348-4298-92ad-a93a70da08cd_0(96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188" Netns:"/var/run/netns/0d4114f6-56ab-462f-bee6-df5e56aade02" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188;K8S_POD_UID=17591eff-a348-4298-92ad-a93a70da08cd" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/17591eff-a348-4298-92ad-a93a70da08cd]: expected pod UID "17591eff-a348-4298-92ad-a93a70da08cd" but got "4002bd9a-1957-499c-b077-3f1eee6b239e" from Kube API Oct 02 11:16:26 crc kubenswrapper[4751]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 02 11:16:26 crc kubenswrapper[4751]: > Oct 02 11:16:26 crc kubenswrapper[4751]: E1002 11:16:26.682601 4751 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 02 11:16:26 crc kubenswrapper[4751]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_17591eff-a348-4298-92ad-a93a70da08cd_0(96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188" Netns:"/var/run/netns/0d4114f6-56ab-462f-bee6-df5e56aade02" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=96cc706e1ddb3f3258581fb228b8ad036b151c560e601a35648256573a04c188;K8S_POD_UID=17591eff-a348-4298-92ad-a93a70da08cd" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/17591eff-a348-4298-92ad-a93a70da08cd]: expected pod UID "17591eff-a348-4298-92ad-a93a70da08cd" but got "4002bd9a-1957-499c-b077-3f1eee6b239e" from Kube API Oct 02 11:16:26 crc kubenswrapper[4751]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 02 11:16:26 crc kubenswrapper[4751]: > pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.707096 4751 generic.go:334] "Generic (PLEG): container finished" podID="361e9060-8522-477e-b827-e2113097fa95" containerID="f0e4db12ef459dc82be3e2b1ce220327a9364585e32feab449e7501cad2abc21" exitCode=0 Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.707417 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" event={"ID":"361e9060-8522-477e-b827-e2113097fa95","Type":"ContainerDied","Data":"f0e4db12ef459dc82be3e2b1ce220327a9364585e32feab449e7501cad2abc21"} Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.707473 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" event={"ID":"361e9060-8522-477e-b827-e2113097fa95","Type":"ContainerStarted","Data":"732f0558d91e28c1b6ff61a98720f98a1a8705d4ffa3e17d00e9e833794324e0"} Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.711274 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.711717 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerStarted","Data":"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d"} Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.711777 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerStarted","Data":"3db1ca90e6f43842f5c7034fafea1da20ecc0d8e00126cd0d017622e384373db"} Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.713095 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.713229 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.713305 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config-secret\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.713336 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz8qn\" (UniqueName: \"kubernetes.io/projected/4002bd9a-1957-499c-b077-3f1eee6b239e-kube-api-access-vz8qn\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.716540 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.720961 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-openstack-config-secret\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.724884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4002bd9a-1957-499c-b077-3f1eee6b239e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.733239 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.734268 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz8qn\" (UniqueName: \"kubernetes.io/projected/4002bd9a-1957-499c-b077-3f1eee6b239e-kube-api-access-vz8qn\") pod \"openstackclient\" (UID: \"4002bd9a-1957-499c-b077-3f1eee6b239e\") " pod="openstack/openstackclient" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.735751 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="17591eff-a348-4298-92ad-a93a70da08cd" podUID="4002bd9a-1957-499c-b077-3f1eee6b239e" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.814834 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config\") pod \"17591eff-a348-4298-92ad-a93a70da08cd\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.814887 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret\") pod \"17591eff-a348-4298-92ad-a93a70da08cd\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.814975 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle\") pod \"17591eff-a348-4298-92ad-a93a70da08cd\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.815034 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xngm\" (UniqueName: \"kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm\") pod \"17591eff-a348-4298-92ad-a93a70da08cd\" (UID: \"17591eff-a348-4298-92ad-a93a70da08cd\") " Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.816315 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "17591eff-a348-4298-92ad-a93a70da08cd" (UID: "17591eff-a348-4298-92ad-a93a70da08cd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.839795 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17591eff-a348-4298-92ad-a93a70da08cd" (UID: "17591eff-a348-4298-92ad-a93a70da08cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.839813 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "17591eff-a348-4298-92ad-a93a70da08cd" (UID: "17591eff-a348-4298-92ad-a93a70da08cd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.842751 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm" (OuterVolumeSpecName: "kube-api-access-2xngm") pod "17591eff-a348-4298-92ad-a93a70da08cd" (UID: "17591eff-a348-4298-92ad-a93a70da08cd"). InnerVolumeSpecName "kube-api-access-2xngm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.918327 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.918364 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17591eff-a348-4298-92ad-a93a70da08cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.918373 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xngm\" (UniqueName: \"kubernetes.io/projected/17591eff-a348-4298-92ad-a93a70da08cd-kube-api-access-2xngm\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:26 crc kubenswrapper[4751]: I1002 11:16:26.918384 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17591eff-a348-4298-92ad-a93a70da08cd-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.025481 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.500412 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 11:16:27 crc kubenswrapper[4751]: W1002 11:16:27.509905 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4002bd9a_1957_499c_b077_3f1eee6b239e.slice/crio-be7accacd97faf7e87a0f8236be74ecc989925a3bb95292968a739f700277efe WatchSource:0}: Error finding container be7accacd97faf7e87a0f8236be74ecc989925a3bb95292968a739f700277efe: Status 404 returned error can't find the container with id be7accacd97faf7e87a0f8236be74ecc989925a3bb95292968a739f700277efe Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.561645 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17591eff-a348-4298-92ad-a93a70da08cd" path="/var/lib/kubelet/pods/17591eff-a348-4298-92ad-a93a70da08cd/volumes" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.671737 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.674188 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.689734 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.720564 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerStarted","Data":"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9"} Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.722361 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.725527 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4002bd9a-1957-499c-b077-3f1eee6b239e","Type":"ContainerStarted","Data":"be7accacd97faf7e87a0f8236be74ecc989925a3bb95292968a739f700277efe"} Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.730453 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" event={"ID":"361e9060-8522-477e-b827-e2113097fa95","Type":"ContainerStarted","Data":"d37aef03089a7b26ef327039ddf0e186d0c2f858539bb236a7e52dbdcce724ba"} Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.730599 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.730627 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.745333 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d9c6c568d-bsnqp" podStartSLOduration=2.745305391 podStartE2EDuration="2.745305391s" podCreationTimestamp="2025-10-02 11:16:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:27.741263153 +0000 UTC m=+1469.795489603" watchObservedRunningTime="2025-10-02 11:16:27.745305391 +0000 UTC m=+1469.799531851" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.762351 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" podStartSLOduration=3.762333666 podStartE2EDuration="3.762333666s" podCreationTimestamp="2025-10-02 11:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:27.756987283 +0000 UTC m=+1469.811213753" watchObservedRunningTime="2025-10-02 11:16:27.762333666 +0000 UTC m=+1469.816560136" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.763865 4751 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="17591eff-a348-4298-92ad-a93a70da08cd" podUID="4002bd9a-1957-499c-b077-3f1eee6b239e" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.833932 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.834486 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dshxj\" (UniqueName: \"kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.834550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.937086 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.937274 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dshxj\" (UniqueName: \"kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.937308 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.937838 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.937991 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.958621 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dshxj\" (UniqueName: \"kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj\") pod \"community-operators-w7lj5\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.978751 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b677f5d65-d5g4x"] Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.980741 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.982771 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.982973 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.991454 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b677f5d65-d5g4x"] Oct 02 11:16:27 crc kubenswrapper[4751]: I1002 11:16:27.994553 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140043 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-httpd-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-combined-ca-bundle\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140491 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-internal-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140555 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140576 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-ovndb-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140637 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-public-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.140677 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb2lw\" (UniqueName: \"kubernetes.io/projected/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-kube-api-access-jb2lw\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242337 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-combined-ca-bundle\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242386 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-httpd-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-internal-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242525 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242555 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-ovndb-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.242636 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-public-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.243923 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb2lw\" (UniqueName: \"kubernetes.io/projected/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-kube-api-access-jb2lw\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.251151 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-combined-ca-bundle\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.251448 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-internal-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.252993 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-ovndb-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.255384 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-public-tls-certs\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.255514 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-httpd-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.255963 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-config\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.266136 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb2lw\" (UniqueName: \"kubernetes.io/projected/5bcbea36-e54d-4ea2-8911-6060ca0edf2e-kube-api-access-jb2lw\") pod \"neutron-7b677f5d65-d5g4x\" (UID: \"5bcbea36-e54d-4ea2-8911-6060ca0edf2e\") " pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.308108 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.419995 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.746746 4751 generic.go:334] "Generic (PLEG): container finished" podID="faae59de-da62-46ae-b58e-acb43ada5fec" containerID="2c701b07095f80e46dcb30b2fb96048e33d67f797796f9b03d7ce01770fc0a5b" exitCode=0 Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.746787 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28csr" event={"ID":"faae59de-da62-46ae-b58e-acb43ada5fec","Type":"ContainerDied","Data":"2c701b07095f80e46dcb30b2fb96048e33d67f797796f9b03d7ce01770fc0a5b"} Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.750277 4751 generic.go:334] "Generic (PLEG): container finished" podID="cb8235e6-b608-498e-b353-4197657298ea" containerID="97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847" exitCode=0 Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.750562 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerDied","Data":"97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847"} Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.750783 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerStarted","Data":"ef3f5a0d436838f35efc699d05c16ef75e88ac44b87b51c841f7b8736a3352b9"} Oct 02 11:16:28 crc kubenswrapper[4751]: I1002 11:16:28.821736 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b677f5d65-d5g4x"] Oct 02 11:16:28 crc kubenswrapper[4751]: W1002 11:16:28.826654 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bcbea36_e54d_4ea2_8911_6060ca0edf2e.slice/crio-3d3cd166f3bd69a3a2ea6197cf26592871a712ffb0fb8f90f844d7e05ccdf881 WatchSource:0}: Error finding container 3d3cd166f3bd69a3a2ea6197cf26592871a712ffb0fb8f90f844d7e05ccdf881: Status 404 returned error can't find the container with id 3d3cd166f3bd69a3a2ea6197cf26592871a712ffb0fb8f90f844d7e05ccdf881 Oct 02 11:16:29 crc kubenswrapper[4751]: I1002 11:16:29.762723 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b677f5d65-d5g4x" event={"ID":"5bcbea36-e54d-4ea2-8911-6060ca0edf2e","Type":"ContainerStarted","Data":"05b7b12c84b34c9392d8df89d37b1da89c6478763dd4208b84568905c65020f3"} Oct 02 11:16:29 crc kubenswrapper[4751]: I1002 11:16:29.763379 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b677f5d65-d5g4x" event={"ID":"5bcbea36-e54d-4ea2-8911-6060ca0edf2e","Type":"ContainerStarted","Data":"756b1a886b1a8c8a3748f8ee2d9f7d40297b0ecc6ed1d76d26adb63ab5157389"} Oct 02 11:16:29 crc kubenswrapper[4751]: I1002 11:16:29.763398 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b677f5d65-d5g4x" event={"ID":"5bcbea36-e54d-4ea2-8911-6060ca0edf2e","Type":"ContainerStarted","Data":"3d3cd166f3bd69a3a2ea6197cf26592871a712ffb0fb8f90f844d7e05ccdf881"} Oct 02 11:16:29 crc kubenswrapper[4751]: I1002 11:16:29.803980 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b677f5d65-d5g4x" podStartSLOduration=2.803954573 podStartE2EDuration="2.803954573s" podCreationTimestamp="2025-10-02 11:16:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:29.789192889 +0000 UTC m=+1471.843419359" watchObservedRunningTime="2025-10-02 11:16:29.803954573 +0000 UTC m=+1471.858181023" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.186100 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28csr" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287202 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g4n2\" (UniqueName: \"kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287324 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287399 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287444 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287465 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287562 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle\") pod \"faae59de-da62-46ae-b58e-acb43ada5fec\" (UID: \"faae59de-da62-46ae-b58e-acb43ada5fec\") " Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287711 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.287907 4751 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/faae59de-da62-46ae-b58e-acb43ada5fec-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.294402 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts" (OuterVolumeSpecName: "scripts") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.296342 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.297894 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2" (OuterVolumeSpecName: "kube-api-access-7g4n2") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "kube-api-access-7g4n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.330529 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.347225 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data" (OuterVolumeSpecName: "config-data") pod "faae59de-da62-46ae-b58e-acb43ada5fec" (UID: "faae59de-da62-46ae-b58e-acb43ada5fec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.390627 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.390662 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.390674 4751 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.390687 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faae59de-da62-46ae-b58e-acb43ada5fec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.390770 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g4n2\" (UniqueName: \"kubernetes.io/projected/faae59de-da62-46ae-b58e-acb43ada5fec-kube-api-access-7g4n2\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.781444 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28csr" event={"ID":"faae59de-da62-46ae-b58e-acb43ada5fec","Type":"ContainerDied","Data":"dbaa52622de083b9a0172473b0e1bd284c7187db15d0db42c4d895ddeee300e1"} Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.781845 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbaa52622de083b9a0172473b0e1bd284c7187db15d0db42c4d895ddeee300e1" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.781459 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28csr" Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.785635 4751 generic.go:334] "Generic (PLEG): container finished" podID="cb8235e6-b608-498e-b353-4197657298ea" containerID="19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68" exitCode=0 Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.785876 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerDied","Data":"19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68"} Oct 02 11:16:30 crc kubenswrapper[4751]: I1002 11:16:30.786023 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.177866 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:31 crc kubenswrapper[4751]: E1002 11:16:31.178289 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" containerName="cinder-db-sync" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.178307 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" containerName="cinder-db-sync" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.178511 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" containerName="cinder-db-sync" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.179446 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.182507 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.182894 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.183092 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ctrq8" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.183213 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.209600 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.266204 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.268222 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="dnsmasq-dns" containerID="cri-o://d37aef03089a7b26ef327039ddf0e186d0c2f858539bb236a7e52dbdcce724ba" gracePeriod=10 Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.280937 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.286566 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318639 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318696 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318730 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318791 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318825 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbv5q\" (UniqueName: \"kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.318854 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.363986 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.421821 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.421878 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jht8l\" (UniqueName: \"kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.427622 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.427746 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.427824 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.427860 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428058 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428139 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbv5q\" (UniqueName: \"kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428196 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428263 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428324 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428381 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.428426 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.430830 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.432850 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.435186 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.439884 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.441876 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.444695 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.445211 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.457984 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.476144 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbv5q\" (UniqueName: \"kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q\") pod \"cinder-scheduler-0\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.500620 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536417 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536489 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536585 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536732 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536791 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536827 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536936 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.536972 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.537082 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.537123 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvflz\" (UniqueName: \"kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.537160 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.537493 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.537642 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jht8l\" (UniqueName: \"kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.538236 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.538760 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.539403 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.539956 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.541260 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.600812 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jht8l\" (UniqueName: \"kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l\") pod \"dnsmasq-dns-6bb4fc677f-lfjcn\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639488 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvflz\" (UniqueName: \"kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639574 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639595 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639635 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639660 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639677 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.639716 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.640068 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.648676 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.648778 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.652788 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.653048 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.655120 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.657156 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvflz\" (UniqueName: \"kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz\") pod \"cinder-api-0\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.692808 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.836873 4751 generic.go:334] "Generic (PLEG): container finished" podID="361e9060-8522-477e-b827-e2113097fa95" containerID="d37aef03089a7b26ef327039ddf0e186d0c2f858539bb236a7e52dbdcce724ba" exitCode=0 Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.838033 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" event={"ID":"361e9060-8522-477e-b827-e2113097fa95","Type":"ContainerDied","Data":"d37aef03089a7b26ef327039ddf0e186d0c2f858539bb236a7e52dbdcce724ba"} Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.850619 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.877709 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.946506 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.951364 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.956242 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth8g\" (UniqueName: \"kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.956353 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.956411 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.956443 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0\") pod \"361e9060-8522-477e-b827-e2113097fa95\" (UID: \"361e9060-8522-477e-b827-e2113097fa95\") " Oct 02 11:16:31 crc kubenswrapper[4751]: I1002 11:16:31.964370 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g" (OuterVolumeSpecName: "kube-api-access-zth8g") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "kube-api-access-zth8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.060470 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zth8g\" (UniqueName: \"kubernetes.io/projected/361e9060-8522-477e-b827-e2113097fa95-kube-api-access-zth8g\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.094673 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.101437 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.143210 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config" (OuterVolumeSpecName: "config") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.144797 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.150629 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.157778 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "361e9060-8522-477e-b827-e2113097fa95" (UID: "361e9060-8522-477e-b827-e2113097fa95"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.179357 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.179551 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.179564 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.179574 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.179584 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e9060-8522-477e-b827-e2113097fa95-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.483244 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.563796 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:16:32 crc kubenswrapper[4751]: W1002 11:16:32.610759 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d4e3bbb_dc71_4be4_b355_eb1b3db0c70c.slice/crio-6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb WatchSource:0}: Error finding container 6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb: Status 404 returned error can't find the container with id 6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.889735 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" event={"ID":"189cc954-c9db-462c-a75d-b7825503e11a","Type":"ContainerStarted","Data":"d192b176f499d8793cd758808525231af9143c1cff314c84795e4e760053c1db"} Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.899747 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerStarted","Data":"17f9aa2e7f105d575a73b34ae21682e2040402e1de15fe47237a204b63ef4020"} Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.907994 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" event={"ID":"361e9060-8522-477e-b827-e2113097fa95","Type":"ContainerDied","Data":"732f0558d91e28c1b6ff61a98720f98a1a8705d4ffa3e17d00e9e833794324e0"} Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.908051 4751 scope.go:117] "RemoveContainer" containerID="d37aef03089a7b26ef327039ddf0e186d0c2f858539bb236a7e52dbdcce724ba" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.908233 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-ntlpt" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.915817 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerStarted","Data":"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c"} Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.929522 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerStarted","Data":"6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb"} Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.948984 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w7lj5" podStartSLOduration=3.305596373 podStartE2EDuration="5.948956954s" podCreationTimestamp="2025-10-02 11:16:27 +0000 UTC" firstStartedPulling="2025-10-02 11:16:28.752998659 +0000 UTC m=+1470.807225109" lastFinishedPulling="2025-10-02 11:16:31.39635925 +0000 UTC m=+1473.450585690" observedRunningTime="2025-10-02 11:16:32.943470548 +0000 UTC m=+1474.997697018" watchObservedRunningTime="2025-10-02 11:16:32.948956954 +0000 UTC m=+1475.003183424" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.955857 4751 scope.go:117] "RemoveContainer" containerID="f0e4db12ef459dc82be3e2b1ce220327a9364585e32feab449e7501cad2abc21" Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.972848 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:32 crc kubenswrapper[4751]: I1002 11:16:32.982452 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-ntlpt"] Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.330590 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.332162 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-m8dqc"] Oct 02 11:16:33 crc kubenswrapper[4751]: E1002 11:16:33.332605 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="dnsmasq-dns" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.332624 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="dnsmasq-dns" Oct 02 11:16:33 crc kubenswrapper[4751]: E1002 11:16:33.332672 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="init" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.332680 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="init" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.332890 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="361e9060-8522-477e-b827-e2113097fa95" containerName="dnsmasq-dns" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.333612 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.349711 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m8dqc"] Oct 02 11:16:33 crc kubenswrapper[4751]: I1002 11:16:33.436615 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsvmv\" (UniqueName: \"kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv\") pod \"nova-api-db-create-m8dqc\" (UID: \"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca\") " pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.440545 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-hfxcj"] Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.442469 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.446577 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hfxcj"] Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.538701 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd9nj\" (UniqueName: \"kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj\") pod \"nova-cell0-db-create-hfxcj\" (UID: \"2c468c9e-b55c-4298-95fd-b382f454f5e6\") " pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.538845 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsvmv\" (UniqueName: \"kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv\") pod \"nova-api-db-create-m8dqc\" (UID: \"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca\") " pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.539455 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-g7q7n"] Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.543736 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.549243 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g7q7n"] Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.590989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsvmv\" (UniqueName: \"kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv\") pod \"nova-api-db-create-m8dqc\" (UID: \"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca\") " pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.592317 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="361e9060-8522-477e-b827-e2113097fa95" path="/var/lib/kubelet/pods/361e9060-8522-477e-b827-e2113097fa95/volumes" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.640725 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd9nj\" (UniqueName: \"kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj\") pod \"nova-cell0-db-create-hfxcj\" (UID: \"2c468c9e-b55c-4298-95fd-b382f454f5e6\") " pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.641006 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sn2g\" (UniqueName: \"kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g\") pod \"nova-cell1-db-create-g7q7n\" (UID: \"0af13930-d981-4282-831e-aa071b1e1e85\") " pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.659717 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.666675 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd9nj\" (UniqueName: \"kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj\") pod \"nova-cell0-db-create-hfxcj\" (UID: \"2c468c9e-b55c-4298-95fd-b382f454f5e6\") " pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.745196 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sn2g\" (UniqueName: \"kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g\") pod \"nova-cell1-db-create-g7q7n\" (UID: \"0af13930-d981-4282-831e-aa071b1e1e85\") " pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.767957 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sn2g\" (UniqueName: \"kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g\") pod \"nova-cell1-db-create-g7q7n\" (UID: \"0af13930-d981-4282-831e-aa071b1e1e85\") " pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.780682 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.844116 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.948745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.967776 4751 generic.go:334] "Generic (PLEG): container finished" podID="189cc954-c9db-462c-a75d-b7825503e11a" containerID="018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513" exitCode=0 Oct 02 11:16:34 crc kubenswrapper[4751]: I1002 11:16:33.968069 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" event={"ID":"189cc954-c9db-462c-a75d-b7825503e11a","Type":"ContainerDied","Data":"018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513"} Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.009268 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerStarted","Data":"27df7f2fe900c6633ea2b661390517c862083e6b4723546a632198ee3c9e029b"} Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.011647 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" event={"ID":"189cc954-c9db-462c-a75d-b7825503e11a","Type":"ContainerStarted","Data":"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8"} Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.011920 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.036412 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" podStartSLOduration=4.036388145 podStartE2EDuration="4.036388145s" podCreationTimestamp="2025-10-02 11:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:35.034097714 +0000 UTC m=+1477.088324164" watchObservedRunningTime="2025-10-02 11:16:35.036388145 +0000 UTC m=+1477.090614595" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.250133 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-67bfc4fc59-jsm55"] Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.251697 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.253874 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.255310 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.255994 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.270194 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-67bfc4fc59-jsm55"] Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.406884 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-log-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.406972 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-combined-ca-bundle\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407029 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-etc-swift\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407063 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-run-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407104 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-config-data\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407135 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-internal-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407200 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmm5x\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-kube-api-access-cmm5x\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.407235 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-public-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.420420 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.420935 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-central-agent" containerID="cri-o://2d8c61f970fb9a0daa43635f8d052ed1bb137f09fa033437f8fcafdb62b5fd95" gracePeriod=30 Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.421045 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" containerID="cri-o://516af6a81663d2aac64b770664cf4d914871763212056fc5759773d959c5c60b" gracePeriod=30 Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.421257 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="sg-core" containerID="cri-o://8e8046a4f3c6192b5b6d2afbaba742a19e3e99048ab5f79ef748e1d34fc002d3" gracePeriod=30 Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.421341 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-notification-agent" containerID="cri-o://e6767c6ef027c584acc5f0ac64d9ee81503b1d109b149ec70b48530d228c295b" gracePeriod=30 Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510064 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-internal-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510161 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmm5x\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-kube-api-access-cmm5x\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510211 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-public-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510315 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-log-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510365 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-combined-ca-bundle\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510413 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-etc-swift\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510446 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-run-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.510484 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-config-data\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.512653 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-run-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.512899 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf4cac44-422b-4c94-9275-5f952d606198-log-httpd\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.517489 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-etc-swift\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.522391 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-public-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.522676 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-config-data\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.524968 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-combined-ca-bundle\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.526336 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": read tcp 10.217.0.2:44418->10.217.0.155:3000: read: connection reset by peer" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.531407 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf4cac44-422b-4c94-9275-5f952d606198-internal-tls-certs\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.548183 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m8dqc"] Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.555565 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmm5x\" (UniqueName: \"kubernetes.io/projected/bf4cac44-422b-4c94-9275-5f952d606198-kube-api-access-cmm5x\") pod \"swift-proxy-67bfc4fc59-jsm55\" (UID: \"bf4cac44-422b-4c94-9275-5f952d606198\") " pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.582460 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.598269 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g7q7n"] Oct 02 11:16:35 crc kubenswrapper[4751]: I1002 11:16:35.598302 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hfxcj"] Oct 02 11:16:35 crc kubenswrapper[4751]: W1002 11:16:35.742143 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c468c9e_b55c_4298_95fd_b382f454f5e6.slice/crio-544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed WatchSource:0}: Error finding container 544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed: Status 404 returned error can't find the container with id 544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed Oct 02 11:16:35 crc kubenswrapper[4751]: W1002 11:16:35.744629 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8c5f5c3_f2a5_4b95_9e64_6c3f02db19ca.slice/crio-cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d WatchSource:0}: Error finding container cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d: Status 404 returned error can't find the container with id cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d Oct 02 11:16:35 crc kubenswrapper[4751]: W1002 11:16:35.745774 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0af13930_d981_4282_831e_aa071b1e1e85.slice/crio-f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891 WatchSource:0}: Error finding container f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891: Status 404 returned error can't find the container with id f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.022029 4751 generic.go:334] "Generic (PLEG): container finished" podID="36fc037e-5c58-4270-93eb-870688ef336b" containerID="516af6a81663d2aac64b770664cf4d914871763212056fc5759773d959c5c60b" exitCode=0 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.022996 4751 generic.go:334] "Generic (PLEG): container finished" podID="36fc037e-5c58-4270-93eb-870688ef336b" containerID="8e8046a4f3c6192b5b6d2afbaba742a19e3e99048ab5f79ef748e1d34fc002d3" exitCode=2 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.023061 4751 generic.go:334] "Generic (PLEG): container finished" podID="36fc037e-5c58-4270-93eb-870688ef336b" containerID="2d8c61f970fb9a0daa43635f8d052ed1bb137f09fa033437f8fcafdb62b5fd95" exitCode=0 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.023187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerDied","Data":"516af6a81663d2aac64b770664cf4d914871763212056fc5759773d959c5c60b"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.023335 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerDied","Data":"8e8046a4f3c6192b5b6d2afbaba742a19e3e99048ab5f79ef748e1d34fc002d3"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.023438 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerDied","Data":"2d8c61f970fb9a0daa43635f8d052ed1bb137f09fa033437f8fcafdb62b5fd95"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.026026 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7q7n" event={"ID":"0af13930-d981-4282-831e-aa071b1e1e85","Type":"ContainerStarted","Data":"f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.027487 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerStarted","Data":"735e5ebf324f2df317859ff894db4da9cc4f2cb1ec4fe2641130c89942d11f15"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.027588 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api-log" containerID="cri-o://27df7f2fe900c6633ea2b661390517c862083e6b4723546a632198ee3c9e029b" gracePeriod=30 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.027696 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api" containerID="cri-o://735e5ebf324f2df317859ff894db4da9cc4f2cb1ec4fe2641130c89942d11f15" gracePeriod=30 Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.027808 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.029948 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8dqc" event={"ID":"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca","Type":"ContainerStarted","Data":"cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.032568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hfxcj" event={"ID":"2c468c9e-b55c-4298-95fd-b382f454f5e6","Type":"ContainerStarted","Data":"544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed"} Oct 02 11:16:36 crc kubenswrapper[4751]: I1002 11:16:36.049162 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.049142578 podStartE2EDuration="5.049142578s" podCreationTimestamp="2025-10-02 11:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:36.046785095 +0000 UTC m=+1478.101011545" watchObservedRunningTime="2025-10-02 11:16:36.049142578 +0000 UTC m=+1478.103369028" Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.048450 4751 generic.go:334] "Generic (PLEG): container finished" podID="36fc037e-5c58-4270-93eb-870688ef336b" containerID="e6767c6ef027c584acc5f0ac64d9ee81503b1d109b149ec70b48530d228c295b" exitCode=0 Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.048510 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerDied","Data":"e6767c6ef027c584acc5f0ac64d9ee81503b1d109b149ec70b48530d228c295b"} Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.058388 4751 generic.go:334] "Generic (PLEG): container finished" podID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerID="27df7f2fe900c6633ea2b661390517c862083e6b4723546a632198ee3c9e029b" exitCode=143 Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.058431 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerDied","Data":"27df7f2fe900c6633ea2b661390517c862083e6b4723546a632198ee3c9e029b"} Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.995527 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:37 crc kubenswrapper[4751]: I1002 11:16:37.995576 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:39 crc kubenswrapper[4751]: I1002 11:16:39.047771 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-w7lj5" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" probeResult="failure" output=< Oct 02 11:16:39 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 11:16:39 crc kubenswrapper[4751]: > Oct 02 11:16:39 crc kubenswrapper[4751]: I1002 11:16:39.931036 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": dial tcp 10.217.0.155:3000: connect: connection refused" Oct 02 11:16:41 crc kubenswrapper[4751]: I1002 11:16:41.695305 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:16:41 crc kubenswrapper[4751]: I1002 11:16:41.758787 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:16:41 crc kubenswrapper[4751]: I1002 11:16:41.759124 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="dnsmasq-dns" containerID="cri-o://6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315" gracePeriod=10 Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.015005 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.129057 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hfxcj" event={"ID":"2c468c9e-b55c-4298-95fd-b382f454f5e6","Type":"ContainerStarted","Data":"d069b51c33c4aaa6cea9bae8af2f121406780a4b63d32970584f8f09a323e1b1"} Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.136983 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7q7n" event={"ID":"0af13930-d981-4282-831e-aa071b1e1e85","Type":"ContainerStarted","Data":"c18d58b46df5495cfd6327a660967ab882759dc5023dd747eef88056e9093a89"} Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.140209 4751 generic.go:334] "Generic (PLEG): container finished" podID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerID="6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315" exitCode=0 Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.140290 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" event={"ID":"6fbe665b-0b68-4c7c-8e45-51685cee78cd","Type":"ContainerDied","Data":"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315"} Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.140323 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" event={"ID":"6fbe665b-0b68-4c7c-8e45-51685cee78cd","Type":"ContainerDied","Data":"a6ca6ccffb381248560cb13a66245795a2625ceecdecfa13cb7d750c4e95ccc0"} Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.140330 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-29wnq" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.140343 4751 scope.go:117] "RemoveContainer" containerID="6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.144215 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8dqc" event={"ID":"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca","Type":"ContainerStarted","Data":"c68f013e7bfc73f4d5edc6719b962c1efc55d54c714c65d771449a895b93e28c"} Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.151039 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168059 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168117 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb7rf\" (UniqueName: \"kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168156 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168260 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168354 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.168510 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config\") pod \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\" (UID: \"6fbe665b-0b68-4c7c-8e45-51685cee78cd\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.180448 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf" (OuterVolumeSpecName: "kube-api-access-bb7rf") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "kube-api-access-bb7rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.224955 4751 scope.go:117] "RemoveContainer" containerID="077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.227198 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.266354 4751 scope.go:117] "RemoveContainer" containerID="6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315" Oct 02 11:16:43 crc kubenswrapper[4751]: E1002 11:16:43.267345 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315\": container with ID starting with 6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315 not found: ID does not exist" containerID="6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.267401 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315"} err="failed to get container status \"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315\": rpc error: code = NotFound desc = could not find container \"6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315\": container with ID starting with 6eee4b603459f0352cf8709b113c0effdf994f08c758d40a63d54c0066172315 not found: ID does not exist" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.267472 4751 scope.go:117] "RemoveContainer" containerID="077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251" Oct 02 11:16:43 crc kubenswrapper[4751]: E1002 11:16:43.269971 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251\": container with ID starting with 077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251 not found: ID does not exist" containerID="077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270006 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251"} err="failed to get container status \"077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251\": rpc error: code = NotFound desc = could not find container \"077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251\": container with ID starting with 077c99228a97d11f37f69c0094f437d9196dcb610eeaac9e8688ad02d18e6251 not found: ID does not exist" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270392 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270568 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77sbx\" (UniqueName: \"kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270645 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270679 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270713 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270809 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.270921 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd\") pod \"36fc037e-5c58-4270-93eb-870688ef336b\" (UID: \"36fc037e-5c58-4270-93eb-870688ef336b\") " Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.271665 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb7rf\" (UniqueName: \"kubernetes.io/projected/6fbe665b-0b68-4c7c-8e45-51685cee78cd-kube-api-access-bb7rf\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.271692 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.272157 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.272957 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.277417 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config" (OuterVolumeSpecName: "config") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.277461 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts" (OuterVolumeSpecName: "scripts") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.282955 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.291786 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.297268 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6fbe665b-0b68-4c7c-8e45-51685cee78cd" (UID: "6fbe665b-0b68-4c7c-8e45-51685cee78cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.301263 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx" (OuterVolumeSpecName: "kube-api-access-77sbx") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "kube-api-access-77sbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.322313 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.330790 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-758c87b4db-9d6vw" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.330919 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374664 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374696 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77sbx\" (UniqueName: \"kubernetes.io/projected/36fc037e-5c58-4270-93eb-870688ef336b-kube-api-access-77sbx\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374707 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374716 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374727 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374736 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374745 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374754 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fbe665b-0b68-4c7c-8e45-51685cee78cd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.374762 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36fc037e-5c58-4270-93eb-870688ef336b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.392479 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.433949 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data" (OuterVolumeSpecName: "config-data") pod "36fc037e-5c58-4270-93eb-870688ef336b" (UID: "36fc037e-5c58-4270-93eb-870688ef336b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.476813 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.477053 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fc037e-5c58-4270-93eb-870688ef336b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.537910 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.543971 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-29wnq"] Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.563237 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" path="/var/lib/kubelet/pods/6fbe665b-0b68-4c7c-8e45-51685cee78cd/volumes" Oct 02 11:16:43 crc kubenswrapper[4751]: I1002 11:16:43.603985 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-67bfc4fc59-jsm55"] Oct 02 11:16:43 crc kubenswrapper[4751]: W1002 11:16:43.605518 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf4cac44_422b_4c94_9275_5f952d606198.slice/crio-cdb8b1d99cc4a6ff9582b3927c8a95ae0d9f325dd6cd85a19a309ab40eaa4d61 WatchSource:0}: Error finding container cdb8b1d99cc4a6ff9582b3927c8a95ae0d9f325dd6cd85a19a309ab40eaa4d61: Status 404 returned error can't find the container with id cdb8b1d99cc4a6ff9582b3927c8a95ae0d9f325dd6cd85a19a309ab40eaa4d61 Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.163955 4751 generic.go:334] "Generic (PLEG): container finished" podID="0af13930-d981-4282-831e-aa071b1e1e85" containerID="c18d58b46df5495cfd6327a660967ab882759dc5023dd747eef88056e9093a89" exitCode=0 Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.164026 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7q7n" event={"ID":"0af13930-d981-4282-831e-aa071b1e1e85","Type":"ContainerDied","Data":"c18d58b46df5495cfd6327a660967ab882759dc5023dd747eef88056e9093a89"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.167200 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67bfc4fc59-jsm55" event={"ID":"bf4cac44-422b-4c94-9275-5f952d606198","Type":"ContainerStarted","Data":"9787da43e8728cba726914c1d2d4c8942b5c09d5739b9d560fedabc93ed82c1d"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.167240 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67bfc4fc59-jsm55" event={"ID":"bf4cac44-422b-4c94-9275-5f952d606198","Type":"ContainerStarted","Data":"cdb8b1d99cc4a6ff9582b3927c8a95ae0d9f325dd6cd85a19a309ab40eaa4d61"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.169879 4751 generic.go:334] "Generic (PLEG): container finished" podID="d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" containerID="c68f013e7bfc73f4d5edc6719b962c1efc55d54c714c65d771449a895b93e28c" exitCode=0 Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.169933 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8dqc" event={"ID":"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca","Type":"ContainerDied","Data":"c68f013e7bfc73f4d5edc6719b962c1efc55d54c714c65d771449a895b93e28c"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.171229 4751 generic.go:334] "Generic (PLEG): container finished" podID="2c468c9e-b55c-4298-95fd-b382f454f5e6" containerID="d069b51c33c4aaa6cea9bae8af2f121406780a4b63d32970584f8f09a323e1b1" exitCode=0 Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.171379 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hfxcj" event={"ID":"2c468c9e-b55c-4298-95fd-b382f454f5e6","Type":"ContainerDied","Data":"d069b51c33c4aaa6cea9bae8af2f121406780a4b63d32970584f8f09a323e1b1"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.182388 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36fc037e-5c58-4270-93eb-870688ef336b","Type":"ContainerDied","Data":"3756ebb7f16c937287a2186b0644bef120fff85f837fa23d716a59d3215738dd"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.182446 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.182469 4751 scope.go:117] "RemoveContainer" containerID="516af6a81663d2aac64b770664cf4d914871763212056fc5759773d959c5c60b" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.184030 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerStarted","Data":"eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b"} Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.229585 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.304361 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.338276 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.353678 4751 scope.go:117] "RemoveContainer" containerID="8e8046a4f3c6192b5b6d2afbaba742a19e3e99048ab5f79ef748e1d34fc002d3" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.373645 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374089 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="sg-core" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374106 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="sg-core" Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374122 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374130 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374144 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-notification-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374151 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-notification-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374182 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="init" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374190 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="init" Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374221 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-central-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374230 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-central-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: E1002 11:16:44.374250 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="dnsmasq-dns" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374257 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="dnsmasq-dns" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374739 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="proxy-httpd" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374755 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="sg-core" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374767 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbe665b-0b68-4c7c-8e45-51685cee78cd" containerName="dnsmasq-dns" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374780 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-notification-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.374799 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fc037e-5c58-4270-93eb-870688ef336b" containerName="ceilometer-central-agent" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.376738 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.377295 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.380409 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.380810 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501083 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501316 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gldq4\" (UniqueName: \"kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501359 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501415 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501441 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501458 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.501499 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.551716 4751 scope.go:117] "RemoveContainer" containerID="e6767c6ef027c584acc5f0ac64d9ee81503b1d109b149ec70b48530d228c295b" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.602951 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603058 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gldq4\" (UniqueName: \"kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603110 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603206 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603247 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603277 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603331 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603954 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.603989 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.606896 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.608422 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.610845 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.619976 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.644430 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gldq4\" (UniqueName: \"kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4\") pod \"ceilometer-0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.694391 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:44 crc kubenswrapper[4751]: I1002 11:16:44.694934 4751 scope.go:117] "RemoveContainer" containerID="2d8c61f970fb9a0daa43635f8d052ed1bb137f09fa033437f8fcafdb62b5fd95" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.195927 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67bfc4fc59-jsm55" event={"ID":"bf4cac44-422b-4c94-9275-5f952d606198","Type":"ContainerStarted","Data":"91b85c57a7a6f22a80fc2788b660597a54e9cb14955bc8c7fcbaed09114576e6"} Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.196017 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.196045 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.197752 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4002bd9a-1957-499c-b077-3f1eee6b239e","Type":"ContainerStarted","Data":"8750204915dd4a7d73cb35ea891f080138d988d21d129699c248dc7307d8070f"} Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.204735 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerStarted","Data":"0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467"} Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.255910 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-67bfc4fc59-jsm55" podStartSLOduration=10.255877112 podStartE2EDuration="10.255877112s" podCreationTimestamp="2025-10-02 11:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:45.23295981 +0000 UTC m=+1487.287186270" watchObservedRunningTime="2025-10-02 11:16:45.255877112 +0000 UTC m=+1487.310103572" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.261547 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.879606884 podStartE2EDuration="14.261524453s" podCreationTimestamp="2025-10-02 11:16:31 +0000 UTC" firstStartedPulling="2025-10-02 11:16:32.23428964 +0000 UTC m=+1474.288516080" lastFinishedPulling="2025-10-02 11:16:42.616207199 +0000 UTC m=+1484.670433649" observedRunningTime="2025-10-02 11:16:45.25542198 +0000 UTC m=+1487.309648430" watchObservedRunningTime="2025-10-02 11:16:45.261524453 +0000 UTC m=+1487.315750903" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.277587 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.286504 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.427573068 podStartE2EDuration="19.28648081s" podCreationTimestamp="2025-10-02 11:16:26 +0000 UTC" firstStartedPulling="2025-10-02 11:16:27.512512198 +0000 UTC m=+1469.566738648" lastFinishedPulling="2025-10-02 11:16:44.37141994 +0000 UTC m=+1486.425646390" observedRunningTime="2025-10-02 11:16:45.275506197 +0000 UTC m=+1487.329732667" watchObservedRunningTime="2025-10-02 11:16:45.28648081 +0000 UTC m=+1487.340707280" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.356717 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.573823 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36fc037e-5c58-4270-93eb-870688ef336b" path="/var/lib/kubelet/pods/36fc037e-5c58-4270-93eb-870688ef336b/volumes" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.623612 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.726428 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsvmv\" (UniqueName: \"kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv\") pod \"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca\" (UID: \"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca\") " Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.735469 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv" (OuterVolumeSpecName: "kube-api-access-vsvmv") pod "d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" (UID: "d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca"). InnerVolumeSpecName "kube-api-access-vsvmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.764983 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.790447 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.829137 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsvmv\" (UniqueName: \"kubernetes.io/projected/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca-kube-api-access-vsvmv\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.930103 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sn2g\" (UniqueName: \"kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g\") pod \"0af13930-d981-4282-831e-aa071b1e1e85\" (UID: \"0af13930-d981-4282-831e-aa071b1e1e85\") " Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.930441 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd9nj\" (UniqueName: \"kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj\") pod \"2c468c9e-b55c-4298-95fd-b382f454f5e6\" (UID: \"2c468c9e-b55c-4298-95fd-b382f454f5e6\") " Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.935112 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj" (OuterVolumeSpecName: "kube-api-access-nd9nj") pod "2c468c9e-b55c-4298-95fd-b382f454f5e6" (UID: "2c468c9e-b55c-4298-95fd-b382f454f5e6"). InnerVolumeSpecName "kube-api-access-nd9nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:45 crc kubenswrapper[4751]: I1002 11:16:45.936123 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g" (OuterVolumeSpecName: "kube-api-access-8sn2g") pod "0af13930-d981-4282-831e-aa071b1e1e85" (UID: "0af13930-d981-4282-831e-aa071b1e1e85"). InnerVolumeSpecName "kube-api-access-8sn2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.032813 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd9nj\" (UniqueName: \"kubernetes.io/projected/2c468c9e-b55c-4298-95fd-b382f454f5e6-kube-api-access-nd9nj\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.032853 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sn2g\" (UniqueName: \"kubernetes.io/projected/0af13930-d981-4282-831e-aa071b1e1e85-kube-api-access-8sn2g\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.215617 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m8dqc" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.215653 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m8dqc" event={"ID":"d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca","Type":"ContainerDied","Data":"cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d"} Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.215728 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cabd9f14002545782539cacc365eabbcfa54dbc19a9aa119df52890e6f50d79d" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.217518 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hfxcj" event={"ID":"2c468c9e-b55c-4298-95fd-b382f454f5e6","Type":"ContainerDied","Data":"544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed"} Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.217541 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="544f56030aa75e92c7c74a69aaae20c437cd924fed45c25dbefb6579c6cbe9ed" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.217557 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hfxcj" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.219328 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g7q7n" event={"ID":"0af13930-d981-4282-831e-aa071b1e1e85","Type":"ContainerDied","Data":"f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891"} Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.219370 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f19c2b8eeb2899c4715404fb4523be1c9718b9e08d5354cdba92fd3377b59891" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.219475 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g7q7n" Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.221031 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerStarted","Data":"92fbfa4145960503089ae4a1db3d231b17e7cef09bf2a54d0c5bbc7aa196c346"} Oct 02 11:16:46 crc kubenswrapper[4751]: I1002 11:16:46.501304 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 11:16:48 crc kubenswrapper[4751]: I1002 11:16:48.243455 4751 generic.go:334] "Generic (PLEG): container finished" podID="61916e69-5770-4e34-acae-3b8a551df701" containerID="f351c5cafadb1cd76de3df169a58c548a4b6d6e4f9c992ba4f53d1b041ba6cdf" exitCode=137 Oct 02 11:16:48 crc kubenswrapper[4751]: I1002 11:16:48.243822 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerDied","Data":"f351c5cafadb1cd76de3df169a58c548a4b6d6e4f9c992ba4f53d1b041ba6cdf"} Oct 02 11:16:48 crc kubenswrapper[4751]: I1002 11:16:48.989006 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.049876 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-w7lj5" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" probeResult="failure" output=< Oct 02 11:16:49 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 11:16:49 crc kubenswrapper[4751]: > Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.084925 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085030 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085061 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085097 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085123 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085205 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085235 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj5vq\" (UniqueName: \"kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq\") pod \"61916e69-5770-4e34-acae-3b8a551df701\" (UID: \"61916e69-5770-4e34-acae-3b8a551df701\") " Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.085713 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs" (OuterVolumeSpecName: "logs") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.090438 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq" (OuterVolumeSpecName: "kube-api-access-vj5vq") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "kube-api-access-vj5vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.091901 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.113569 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts" (OuterVolumeSpecName: "scripts") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.118616 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.125075 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data" (OuterVolumeSpecName: "config-data") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.141217 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "61916e69-5770-4e34-acae-3b8a551df701" (UID: "61916e69-5770-4e34-acae-3b8a551df701"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187464 4751 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187509 4751 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187522 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187532 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj5vq\" (UniqueName: \"kubernetes.io/projected/61916e69-5770-4e34-acae-3b8a551df701-kube-api-access-vj5vq\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187547 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61916e69-5770-4e34-acae-3b8a551df701-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187558 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61916e69-5770-4e34-acae-3b8a551df701-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.187567 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61916e69-5770-4e34-acae-3b8a551df701-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.261425 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerStarted","Data":"02282363063a388016092e8aabf717da1265c170aafb6228fb331df9b5c66fa5"} Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.264317 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-758c87b4db-9d6vw" event={"ID":"61916e69-5770-4e34-acae-3b8a551df701","Type":"ContainerDied","Data":"cf3033c6202b89b505b55e5ea2f831594eda5eaf8f1ddca02e35058cf984b123"} Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.264367 4751 scope.go:117] "RemoveContainer" containerID="d56322dba3fe24647b339aa90dbabe8be5b32654c39498af79c501dcd44d87ac" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.264547 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-758c87b4db-9d6vw" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.300694 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.308987 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-758c87b4db-9d6vw"] Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.425242 4751 scope.go:117] "RemoveContainer" containerID="f351c5cafadb1cd76de3df169a58c548a4b6d6e4f9c992ba4f53d1b041ba6cdf" Oct 02 11:16:49 crc kubenswrapper[4751]: I1002 11:16:49.562667 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61916e69-5770-4e34-acae-3b8a551df701" path="/var/lib/kubelet/pods/61916e69-5770-4e34-acae-3b8a551df701/volumes" Oct 02 11:16:50 crc kubenswrapper[4751]: I1002 11:16:50.597771 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:50 crc kubenswrapper[4751]: I1002 11:16:50.600026 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-67bfc4fc59-jsm55" Oct 02 11:16:51 crc kubenswrapper[4751]: I1002 11:16:51.284750 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerStarted","Data":"d57fcd7f21369942e7a005168f6f64148a7af65dd2b6ace21d5fc414e3ed8ca4"} Oct 02 11:16:52 crc kubenswrapper[4751]: I1002 11:16:52.256671 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 11:16:52 crc kubenswrapper[4751]: I1002 11:16:52.304666 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:52 crc kubenswrapper[4751]: I1002 11:16:52.305030 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="cinder-scheduler" containerID="cri-o://eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b" gracePeriod=30 Oct 02 11:16:52 crc kubenswrapper[4751]: I1002 11:16:52.305127 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="probe" containerID="cri-o://0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467" gracePeriod=30 Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.307658 4751 generic.go:334] "Generic (PLEG): container finished" podID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerID="0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467" exitCode=0 Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.307691 4751 generic.go:334] "Generic (PLEG): container finished" podID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerID="eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b" exitCode=0 Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.307724 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerDied","Data":"0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467"} Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.307751 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerDied","Data":"eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b"} Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.309351 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerStarted","Data":"a39b5f8280aa5ddb990283bb282f5b70d1a5a60e951eebbd99d4a69f956840df"} Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.310895 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b413b4_15b9_40e6_b291_c335f54c74f8.slice/crio-conmon-eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b413b4_15b9_40e6_b291_c335f54c74f8.slice/crio-0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b413b4_15b9_40e6_b291_c335f54c74f8.slice/crio-conmon-0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.648876 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-551e-account-create-pw5vt"] Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.649530 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649549 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.649569 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649577 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.649594 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af13930-d981-4282-831e-aa071b1e1e85" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649600 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af13930-d981-4282-831e-aa071b1e1e85" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.649610 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c468c9e-b55c-4298-95fd-b382f454f5e6" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649616 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c468c9e-b55c-4298-95fd-b382f454f5e6" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.649636 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon-log" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649642 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon-log" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649835 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af13930-d981-4282-831e-aa071b1e1e85" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649852 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649875 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c468c9e-b55c-4298-95fd-b382f454f5e6" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649884 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" containerName="mariadb-database-create" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.649897 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="61916e69-5770-4e34-acae-3b8a551df701" containerName="horizon-log" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.650491 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.655380 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.662669 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-551e-account-create-pw5vt"] Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.735242 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.796074 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vn8h\" (UniqueName: \"kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h\") pod \"nova-api-551e-account-create-pw5vt\" (UID: \"2319b758-f6a7-4d3f-b1c8-25747580a255\") " pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.844821 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e375-account-create-7d49s"] Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.845212 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="probe" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.845223 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="probe" Oct 02 11:16:53 crc kubenswrapper[4751]: E1002 11:16:53.845256 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="cinder-scheduler" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.845262 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="cinder-scheduler" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.845447 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="cinder-scheduler" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.845464 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" containerName="probe" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.846054 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.851319 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.873316 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e375-account-create-7d49s"] Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900103 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900251 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900293 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbv5q\" (UniqueName: \"kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900335 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900507 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900557 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data\") pod \"51b413b4-15b9-40e6-b291-c335f54c74f8\" (UID: \"51b413b4-15b9-40e6-b291-c335f54c74f8\") " Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.900922 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vn8h\" (UniqueName: \"kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h\") pod \"nova-api-551e-account-create-pw5vt\" (UID: \"2319b758-f6a7-4d3f-b1c8-25747580a255\") " pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.907598 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.908333 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts" (OuterVolumeSpecName: "scripts") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.926305 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.931487 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vn8h\" (UniqueName: \"kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h\") pod \"nova-api-551e-account-create-pw5vt\" (UID: \"2319b758-f6a7-4d3f-b1c8-25747580a255\") " pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:53 crc kubenswrapper[4751]: I1002 11:16:53.945489 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q" (OuterVolumeSpecName: "kube-api-access-mbv5q") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "kube-api-access-mbv5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.000351 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002358 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckps\" (UniqueName: \"kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps\") pod \"nova-cell0-e375-account-create-7d49s\" (UID: \"7e8d5277-79db-4340-a8fa-66b661abd20a\") " pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002508 4751 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b413b4-15b9-40e6-b291-c335f54c74f8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002519 4751 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002529 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002537 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbv5q\" (UniqueName: \"kubernetes.io/projected/51b413b4-15b9-40e6-b291-c335f54c74f8-kube-api-access-mbv5q\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.002548 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.038521 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a8c7-account-create-cxz7f"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.039725 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.042901 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.048487 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.060159 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a8c7-account-create-cxz7f"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.100920 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data" (OuterVolumeSpecName: "config-data") pod "51b413b4-15b9-40e6-b291-c335f54c74f8" (UID: "51b413b4-15b9-40e6-b291-c335f54c74f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.110087 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ckps\" (UniqueName: \"kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps\") pod \"nova-cell0-e375-account-create-7d49s\" (UID: \"7e8d5277-79db-4340-a8fa-66b661abd20a\") " pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.110396 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b413b4-15b9-40e6-b291-c335f54c74f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.128890 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ckps\" (UniqueName: \"kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps\") pod \"nova-cell0-e375-account-create-7d49s\" (UID: \"7e8d5277-79db-4340-a8fa-66b661abd20a\") " pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.212111 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m6kx\" (UniqueName: \"kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx\") pod \"nova-cell1-a8c7-account-create-cxz7f\" (UID: \"939a3dba-d15a-4257-987d-c931791cf416\") " pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.296681 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.313540 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m6kx\" (UniqueName: \"kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx\") pod \"nova-cell1-a8c7-account-create-cxz7f\" (UID: \"939a3dba-d15a-4257-987d-c931791cf416\") " pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.322522 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerStarted","Data":"79639681f5906df8310cf19b00d539d472cd4683a625e8578fbe15110312fb26"} Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.322735 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-central-agent" containerID="cri-o://02282363063a388016092e8aabf717da1265c170aafb6228fb331df9b5c66fa5" gracePeriod=30 Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.323074 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.323421 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="proxy-httpd" containerID="cri-o://79639681f5906df8310cf19b00d539d472cd4683a625e8578fbe15110312fb26" gracePeriod=30 Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.323483 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="sg-core" containerID="cri-o://a39b5f8280aa5ddb990283bb282f5b70d1a5a60e951eebbd99d4a69f956840df" gracePeriod=30 Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.323532 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-notification-agent" containerID="cri-o://d57fcd7f21369942e7a005168f6f64148a7af65dd2b6ace21d5fc414e3ed8ca4" gracePeriod=30 Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.366947 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"51b413b4-15b9-40e6-b291-c335f54c74f8","Type":"ContainerDied","Data":"17f9aa2e7f105d575a73b34ae21682e2040402e1de15fe47237a204b63ef4020"} Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.367002 4751 scope.go:117] "RemoveContainer" containerID="0dd8e58c9eacc845ab9d055465a2248711c1509c419f09e69182bae053ee3467" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.367185 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.375187 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m6kx\" (UniqueName: \"kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx\") pod \"nova-cell1-a8c7-account-create-cxz7f\" (UID: \"939a3dba-d15a-4257-987d-c931791cf416\") " pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.406343 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.889057619 podStartE2EDuration="10.40629171s" podCreationTimestamp="2025-10-02 11:16:44 +0000 UTC" firstStartedPulling="2025-10-02 11:16:45.292574103 +0000 UTC m=+1487.346800553" lastFinishedPulling="2025-10-02 11:16:53.809808194 +0000 UTC m=+1495.864034644" observedRunningTime="2025-10-02 11:16:54.356659823 +0000 UTC m=+1496.410886283" watchObservedRunningTime="2025-10-02 11:16:54.40629171 +0000 UTC m=+1496.460518170" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.415700 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.432513 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.448228 4751 scope.go:117] "RemoveContainer" containerID="eae03496095e7475197f1e026b12951fd998ea3d976befdfd94c1592cd8f575b" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.466566 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.469412 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.472397 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.475558 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.551152 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-551e-account-create-pw5vt"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.571326 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.619984 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.620032 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz57w\" (UniqueName: \"kubernetes.io/projected/9f7323c9-33fe-466b-9544-85a75f09938d-kube-api-access-fz57w\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.620065 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.620138 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.620158 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.620200 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f7323c9-33fe-466b-9544-85a75f09938d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.667295 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.721763 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.721815 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.721858 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f7323c9-33fe-466b-9544-85a75f09938d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.721946 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.721990 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz57w\" (UniqueName: \"kubernetes.io/projected/9f7323c9-33fe-466b-9544-85a75f09938d-kube-api-access-fz57w\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.722042 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.722636 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f7323c9-33fe-466b-9544-85a75f09938d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.735230 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.735231 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.735943 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.737452 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f7323c9-33fe-466b-9544-85a75f09938d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.738860 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz57w\" (UniqueName: \"kubernetes.io/projected/9f7323c9-33fe-466b-9544-85a75f09938d-kube-api-access-fz57w\") pod \"cinder-scheduler-0\" (UID: \"9f7323c9-33fe-466b-9544-85a75f09938d\") " pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.804803 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.845435 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e375-account-create-7d49s"] Oct 02 11:16:54 crc kubenswrapper[4751]: I1002 11:16:54.975897 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6d856cc8dd-n7jjw" Oct 02 11:16:55 crc kubenswrapper[4751]: W1002 11:16:55.177580 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod939a3dba_d15a_4257_987d_c931791cf416.slice/crio-dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6 WatchSource:0}: Error finding container dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6: Status 404 returned error can't find the container with id dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.178814 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a8c7-account-create-cxz7f"] Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.328710 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 11:16:55 crc kubenswrapper[4751]: W1002 11:16:55.328765 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f7323c9_33fe_466b_9544_85a75f09938d.slice/crio-fdc1290bbe91ed07a4e20edd660aef34254f092e20e494acd39967bdf5183eb2 WatchSource:0}: Error finding container fdc1290bbe91ed07a4e20edd660aef34254f092e20e494acd39967bdf5183eb2: Status 404 returned error can't find the container with id fdc1290bbe91ed07a4e20edd660aef34254f092e20e494acd39967bdf5183eb2 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.379232 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" event={"ID":"939a3dba-d15a-4257-987d-c931791cf416","Type":"ContainerStarted","Data":"dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.381151 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9f7323c9-33fe-466b-9544-85a75f09938d","Type":"ContainerStarted","Data":"fdc1290bbe91ed07a4e20edd660aef34254f092e20e494acd39967bdf5183eb2"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.384272 4751 generic.go:334] "Generic (PLEG): container finished" podID="2319b758-f6a7-4d3f-b1c8-25747580a255" containerID="18f100a45c6dd9cc4f685ce65e2537cb77b6f5f955580996c7c1d1f3fd986a56" exitCode=0 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.384378 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-551e-account-create-pw5vt" event={"ID":"2319b758-f6a7-4d3f-b1c8-25747580a255","Type":"ContainerDied","Data":"18f100a45c6dd9cc4f685ce65e2537cb77b6f5f955580996c7c1d1f3fd986a56"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.384417 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-551e-account-create-pw5vt" event={"ID":"2319b758-f6a7-4d3f-b1c8-25747580a255","Type":"ContainerStarted","Data":"47c24e3d6bdbc1ee3adbb0a10180873558f5e4041e6d8e27bd9fb85a840a8d3a"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387106 4751 generic.go:334] "Generic (PLEG): container finished" podID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerID="79639681f5906df8310cf19b00d539d472cd4683a625e8578fbe15110312fb26" exitCode=0 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387133 4751 generic.go:334] "Generic (PLEG): container finished" podID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerID="a39b5f8280aa5ddb990283bb282f5b70d1a5a60e951eebbd99d4a69f956840df" exitCode=2 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387140 4751 generic.go:334] "Generic (PLEG): container finished" podID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerID="d57fcd7f21369942e7a005168f6f64148a7af65dd2b6ace21d5fc414e3ed8ca4" exitCode=0 Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387209 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerDied","Data":"79639681f5906df8310cf19b00d539d472cd4683a625e8578fbe15110312fb26"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387272 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerDied","Data":"a39b5f8280aa5ddb990283bb282f5b70d1a5a60e951eebbd99d4a69f956840df"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.387288 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerDied","Data":"d57fcd7f21369942e7a005168f6f64148a7af65dd2b6ace21d5fc414e3ed8ca4"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.390916 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e375-account-create-7d49s" event={"ID":"7e8d5277-79db-4340-a8fa-66b661abd20a","Type":"ContainerStarted","Data":"aacb5b67735a8ffbe6b2aabe3a522c641bf07481cfc35b1f438482ec87a91f8e"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.390966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e375-account-create-7d49s" event={"ID":"7e8d5277-79db-4340-a8fa-66b661abd20a","Type":"ContainerStarted","Data":"0dd65c7c9c28e70ec111784190201f58059caf617857cf9247967b026a91ebad"} Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.421730 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-e375-account-create-7d49s" podStartSLOduration=2.421704014 podStartE2EDuration="2.421704014s" podCreationTimestamp="2025-10-02 11:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:55.415643752 +0000 UTC m=+1497.469870212" watchObservedRunningTime="2025-10-02 11:16:55.421704014 +0000 UTC m=+1497.475930484" Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.472634 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:16:55 crc kubenswrapper[4751]: I1002 11:16:55.587298 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b413b4-15b9-40e6-b291-c335f54c74f8" path="/var/lib/kubelet/pods/51b413b4-15b9-40e6-b291-c335f54c74f8/volumes" Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.408311 4751 generic.go:334] "Generic (PLEG): container finished" podID="7e8d5277-79db-4340-a8fa-66b661abd20a" containerID="aacb5b67735a8ffbe6b2aabe3a522c641bf07481cfc35b1f438482ec87a91f8e" exitCode=0 Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.408389 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e375-account-create-7d49s" event={"ID":"7e8d5277-79db-4340-a8fa-66b661abd20a","Type":"ContainerDied","Data":"aacb5b67735a8ffbe6b2aabe3a522c641bf07481cfc35b1f438482ec87a91f8e"} Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.415098 4751 generic.go:334] "Generic (PLEG): container finished" podID="939a3dba-d15a-4257-987d-c931791cf416" containerID="6e4f9dc08829d521d9dca22ef606609231cc5b63de5bc0a108715570307c7fa0" exitCode=0 Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.415210 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" event={"ID":"939a3dba-d15a-4257-987d-c931791cf416","Type":"ContainerDied","Data":"6e4f9dc08829d521d9dca22ef606609231cc5b63de5bc0a108715570307c7fa0"} Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.416823 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9f7323c9-33fe-466b-9544-85a75f09938d","Type":"ContainerStarted","Data":"1d05731f042423c7c2c681f073ba181cf67f4c3d9cdbff752a041de2589dd41f"} Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.810765 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.869347 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vn8h\" (UniqueName: \"kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h\") pod \"2319b758-f6a7-4d3f-b1c8-25747580a255\" (UID: \"2319b758-f6a7-4d3f-b1c8-25747580a255\") " Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.885412 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h" (OuterVolumeSpecName: "kube-api-access-5vn8h") pod "2319b758-f6a7-4d3f-b1c8-25747580a255" (UID: "2319b758-f6a7-4d3f-b1c8-25747580a255"). InnerVolumeSpecName "kube-api-access-5vn8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:56 crc kubenswrapper[4751]: I1002 11:16:56.971804 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vn8h\" (UniqueName: \"kubernetes.io/projected/2319b758-f6a7-4d3f-b1c8-25747580a255-kube-api-access-5vn8h\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.427449 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-551e-account-create-pw5vt" event={"ID":"2319b758-f6a7-4d3f-b1c8-25747580a255","Type":"ContainerDied","Data":"47c24e3d6bdbc1ee3adbb0a10180873558f5e4041e6d8e27bd9fb85a840a8d3a"} Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.427500 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47c24e3d6bdbc1ee3adbb0a10180873558f5e4041e6d8e27bd9fb85a840a8d3a" Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.427562 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-551e-account-create-pw5vt" Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.430224 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9f7323c9-33fe-466b-9544-85a75f09938d","Type":"ContainerStarted","Data":"d3af4c0b654b9c8dcdceeccdf8fc468e98d135050f03dd91a0aae5acf3064a20"} Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.462993 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.46297355 podStartE2EDuration="3.46297355s" podCreationTimestamp="2025-10-02 11:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:16:57.45550375 +0000 UTC m=+1499.509730200" watchObservedRunningTime="2025-10-02 11:16:57.46297355 +0000 UTC m=+1499.517200000" Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.964633 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:57 crc kubenswrapper[4751]: I1002 11:16:57.971700 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.045071 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.090482 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckps\" (UniqueName: \"kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps\") pod \"7e8d5277-79db-4340-a8fa-66b661abd20a\" (UID: \"7e8d5277-79db-4340-a8fa-66b661abd20a\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.091045 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m6kx\" (UniqueName: \"kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx\") pod \"939a3dba-d15a-4257-987d-c931791cf416\" (UID: \"939a3dba-d15a-4257-987d-c931791cf416\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.096370 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.096800 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps" (OuterVolumeSpecName: "kube-api-access-2ckps") pod "7e8d5277-79db-4340-a8fa-66b661abd20a" (UID: "7e8d5277-79db-4340-a8fa-66b661abd20a"). InnerVolumeSpecName "kube-api-access-2ckps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.098516 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx" (OuterVolumeSpecName: "kube-api-access-8m6kx") pod "939a3dba-d15a-4257-987d-c931791cf416" (UID: "939a3dba-d15a-4257-987d-c931791cf416"). InnerVolumeSpecName "kube-api-access-8m6kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.193481 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m6kx\" (UniqueName: \"kubernetes.io/projected/939a3dba-d15a-4257-987d-c931791cf416-kube-api-access-8m6kx\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.193530 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ckps\" (UniqueName: \"kubernetes.io/projected/7e8d5277-79db-4340-a8fa-66b661abd20a-kube-api-access-2ckps\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.441672 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.441692 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a8c7-account-create-cxz7f" event={"ID":"939a3dba-d15a-4257-987d-c931791cf416","Type":"ContainerDied","Data":"dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6"} Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.442676 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcc7ef9f9a736383ad3d1ffbe3cacc8579f417a4dbc23433d87c071185abdfe6" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.443312 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b677f5d65-d5g4x" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.448323 4751 generic.go:334] "Generic (PLEG): container finished" podID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerID="02282363063a388016092e8aabf717da1265c170aafb6228fb331df9b5c66fa5" exitCode=0 Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.448387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerDied","Data":"02282363063a388016092e8aabf717da1265c170aafb6228fb331df9b5c66fa5"} Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.454040 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e375-account-create-7d49s" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.454034 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e375-account-create-7d49s" event={"ID":"7e8d5277-79db-4340-a8fa-66b661abd20a","Type":"ContainerDied","Data":"0dd65c7c9c28e70ec111784190201f58059caf617857cf9247967b026a91ebad"} Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.454200 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dd65c7c9c28e70ec111784190201f58059caf617857cf9247967b026a91ebad" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.515653 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.515957 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9c6c568d-bsnqp" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-api" containerID="cri-o://288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d" gracePeriod=30 Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.521263 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9c6c568d-bsnqp" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-httpd" containerID="cri-o://3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9" gracePeriod=30 Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.708613 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807289 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807373 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807478 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gldq4\" (UniqueName: \"kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807624 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807654 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807688 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.807792 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts\") pod \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\" (UID: \"fd6b4b52-0423-4a5f-9026-84aa87f19db0\") " Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.808735 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.809076 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.814423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts" (OuterVolumeSpecName: "scripts") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.815159 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4" (OuterVolumeSpecName: "kube-api-access-gldq4") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "kube-api-access-gldq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.857695 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.876787 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.910439 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.910470 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.910480 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd6b4b52-0423-4a5f-9026-84aa87f19db0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.910489 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.910498 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gldq4\" (UniqueName: \"kubernetes.io/projected/fd6b4b52-0423-4a5f-9026-84aa87f19db0-kube-api-access-gldq4\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.924535 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:58 crc kubenswrapper[4751]: I1002 11:16:58.975504 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data" (OuterVolumeSpecName: "config-data") pod "fd6b4b52-0423-4a5f-9026-84aa87f19db0" (UID: "fd6b4b52-0423-4a5f-9026-84aa87f19db0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.044383 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.044449 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6b4b52-0423-4a5f-9026-84aa87f19db0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.464631 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd6b4b52-0423-4a5f-9026-84aa87f19db0","Type":"ContainerDied","Data":"92fbfa4145960503089ae4a1db3d231b17e7cef09bf2a54d0c5bbc7aa196c346"} Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.464683 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.464696 4751 scope.go:117] "RemoveContainer" containerID="79639681f5906df8310cf19b00d539d472cd4683a625e8578fbe15110312fb26" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.468684 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerID="3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9" exitCode=0 Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.468726 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerDied","Data":"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9"} Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.468944 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w7lj5" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" containerID="cri-o://0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c" gracePeriod=2 Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.499031 4751 scope.go:117] "RemoveContainer" containerID="a39b5f8280aa5ddb990283bb282f5b70d1a5a60e951eebbd99d4a69f956840df" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.507384 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.522240 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.533843 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534312 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-notification-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534330 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-notification-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534339 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="sg-core" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534346 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="sg-core" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534368 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e8d5277-79db-4340-a8fa-66b661abd20a" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534376 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e8d5277-79db-4340-a8fa-66b661abd20a" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534407 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="proxy-httpd" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534416 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="proxy-httpd" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534429 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2319b758-f6a7-4d3f-b1c8-25747580a255" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534436 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2319b758-f6a7-4d3f-b1c8-25747580a255" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534453 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-central-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534460 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-central-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: E1002 11:16:59.534474 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="939a3dba-d15a-4257-987d-c931791cf416" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534480 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="939a3dba-d15a-4257-987d-c931791cf416" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534704 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-central-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534732 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2319b758-f6a7-4d3f-b1c8-25747580a255" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534745 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e8d5277-79db-4340-a8fa-66b661abd20a" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534762 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="ceilometer-notification-agent" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534773 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="sg-core" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534788 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="939a3dba-d15a-4257-987d-c931791cf416" containerName="mariadb-account-create" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.534801 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" containerName="proxy-httpd" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.536810 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.537628 4751 scope.go:117] "RemoveContainer" containerID="d57fcd7f21369942e7a005168f6f64148a7af65dd2b6ace21d5fc414e3ed8ca4" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.539325 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.539480 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.584584 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd6b4b52-0423-4a5f-9026-84aa87f19db0" path="/var/lib/kubelet/pods/fd6b4b52-0423-4a5f-9026-84aa87f19db0/volumes" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.585460 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.603432 4751 scope.go:117] "RemoveContainer" containerID="02282363063a388016092e8aabf717da1265c170aafb6228fb331df9b5c66fa5" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655124 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655193 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655224 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655264 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655301 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655378 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmdp8\" (UniqueName: \"kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.655394 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.756896 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.756950 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.756981 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757012 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757044 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmdp8\" (UniqueName: \"kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757062 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757159 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757655 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.757887 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.761952 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.762033 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.764068 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.764560 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.776827 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmdp8\" (UniqueName: \"kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8\") pod \"ceilometer-0\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " pod="openstack/ceilometer-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.809028 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 11:16:59 crc kubenswrapper[4751]: I1002 11:16:59.889770 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.018233 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.174402 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities\") pod \"cb8235e6-b608-498e-b353-4197657298ea\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.174600 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content\") pod \"cb8235e6-b608-498e-b353-4197657298ea\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.174636 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dshxj\" (UniqueName: \"kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj\") pod \"cb8235e6-b608-498e-b353-4197657298ea\" (UID: \"cb8235e6-b608-498e-b353-4197657298ea\") " Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.174922 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities" (OuterVolumeSpecName: "utilities") pod "cb8235e6-b608-498e-b353-4197657298ea" (UID: "cb8235e6-b608-498e-b353-4197657298ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.175206 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.179460 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj" (OuterVolumeSpecName: "kube-api-access-dshxj") pod "cb8235e6-b608-498e-b353-4197657298ea" (UID: "cb8235e6-b608-498e-b353-4197657298ea"). InnerVolumeSpecName "kube-api-access-dshxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.224839 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb8235e6-b608-498e-b353-4197657298ea" (UID: "cb8235e6-b608-498e-b353-4197657298ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.277387 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb8235e6-b608-498e-b353-4197657298ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.277422 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dshxj\" (UniqueName: \"kubernetes.io/projected/cb8235e6-b608-498e-b353-4197657298ea-kube-api-access-dshxj\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.359996 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.483990 4751 generic.go:334] "Generic (PLEG): container finished" podID="cb8235e6-b608-498e-b353-4197657298ea" containerID="0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c" exitCode=0 Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.484081 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerDied","Data":"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c"} Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.484108 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w7lj5" event={"ID":"cb8235e6-b608-498e-b353-4197657298ea","Type":"ContainerDied","Data":"ef3f5a0d436838f35efc699d05c16ef75e88ac44b87b51c841f7b8736a3352b9"} Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.484126 4751 scope.go:117] "RemoveContainer" containerID="0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.484970 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w7lj5" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.492362 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerStarted","Data":"0c914d790c9d0174f73923a2091d0ada0bde3847ccc0a9a4003e27034860eb15"} Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.526572 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.533743 4751 scope.go:117] "RemoveContainer" containerID="19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.539091 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w7lj5"] Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.560325 4751 scope.go:117] "RemoveContainer" containerID="97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.578718 4751 scope.go:117] "RemoveContainer" containerID="0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c" Oct 02 11:17:00 crc kubenswrapper[4751]: E1002 11:17:00.579254 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c\": container with ID starting with 0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c not found: ID does not exist" containerID="0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.579315 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c"} err="failed to get container status \"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c\": rpc error: code = NotFound desc = could not find container \"0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c\": container with ID starting with 0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c not found: ID does not exist" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.579351 4751 scope.go:117] "RemoveContainer" containerID="19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68" Oct 02 11:17:00 crc kubenswrapper[4751]: E1002 11:17:00.579615 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68\": container with ID starting with 19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68 not found: ID does not exist" containerID="19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.579647 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68"} err="failed to get container status \"19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68\": rpc error: code = NotFound desc = could not find container \"19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68\": container with ID starting with 19b21f38c2bb8554a82160ded8bca92db2a0f3e6a9be98751a766b9b5fcbbd68 not found: ID does not exist" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.579669 4751 scope.go:117] "RemoveContainer" containerID="97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847" Oct 02 11:17:00 crc kubenswrapper[4751]: E1002 11:17:00.580386 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847\": container with ID starting with 97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847 not found: ID does not exist" containerID="97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847" Oct 02 11:17:00 crc kubenswrapper[4751]: I1002 11:17:00.580423 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847"} err="failed to get container status \"97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847\": rpc error: code = NotFound desc = could not find container \"97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847\": container with ID starting with 97361cb48ffe9e9a2bf1bcd9b315d46ebf9f0aa754d668907fd8982240231847 not found: ID does not exist" Oct 02 11:17:01 crc kubenswrapper[4751]: I1002 11:17:01.506343 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerStarted","Data":"0028e9e60d949d513a2eac9c0e58d7495bf81bb81c42320bef3126557c644663"} Oct 02 11:17:01 crc kubenswrapper[4751]: I1002 11:17:01.561672 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb8235e6-b608-498e-b353-4197657298ea" path="/var/lib/kubelet/pods/cb8235e6-b608-498e-b353-4197657298ea/volumes" Oct 02 11:17:02 crc kubenswrapper[4751]: I1002 11:17:02.519187 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerStarted","Data":"300f2889f589fdeea428d3f849bf8c0a27c4571ca76789e371de4ae4dd7b06e4"} Oct 02 11:17:03 crc kubenswrapper[4751]: I1002 11:17:03.528702 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerStarted","Data":"0dc7b6ba193eb578d87cfd0f949ac57008c192ec0ee6dd2c3f88d28db1065830"} Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.133063 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-q7dsc"] Oct 02 11:17:04 crc kubenswrapper[4751]: E1002 11:17:04.133452 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="extract-content" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.133468 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="extract-content" Oct 02 11:17:04 crc kubenswrapper[4751]: E1002 11:17:04.133498 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="extract-utilities" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.133506 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="extract-utilities" Oct 02 11:17:04 crc kubenswrapper[4751]: E1002 11:17:04.133527 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.133534 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.133699 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb8235e6-b608-498e-b353-4197657298ea" containerName="registry-server" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.134257 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.136934 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.137125 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kjgf2" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.137507 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.152752 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-q7dsc"] Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.248016 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.248158 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.248202 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.248219 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g6xl\" (UniqueName: \"kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.350049 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.350112 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.350138 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g6xl\" (UniqueName: \"kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.350157 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.365282 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.365731 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.366569 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.367623 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g6xl\" (UniqueName: \"kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl\") pod \"nova-cell0-conductor-db-sync-q7dsc\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.456816 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.555404 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerStarted","Data":"d4836bf48fcf8388404521646d5ed11cd88ea2dca5632cf5233ffa9237155d8b"} Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.555813 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.917396 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.132081612 podStartE2EDuration="5.917368769s" podCreationTimestamp="2025-10-02 11:16:59 +0000 UTC" firstStartedPulling="2025-10-02 11:17:00.365930341 +0000 UTC m=+1502.420156791" lastFinishedPulling="2025-10-02 11:17:04.151217498 +0000 UTC m=+1506.205443948" observedRunningTime="2025-10-02 11:17:04.584045649 +0000 UTC m=+1506.638272119" watchObservedRunningTime="2025-10-02 11:17:04.917368769 +0000 UTC m=+1506.971595219" Oct 02 11:17:04 crc kubenswrapper[4751]: I1002 11:17:04.923896 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-q7dsc"] Oct 02 11:17:05 crc kubenswrapper[4751]: I1002 11:17:05.004377 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 11:17:05 crc kubenswrapper[4751]: I1002 11:17:05.567107 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" event={"ID":"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d","Type":"ContainerStarted","Data":"583ee1d11b11fd9d350c24ea25f1c758ca320d11f7fbbbb8fc3a4c46594f5aa8"} Oct 02 11:17:06 crc kubenswrapper[4751]: E1002 11:17:06.411318 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.593737 4751 generic.go:334] "Generic (PLEG): container finished" podID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerID="735e5ebf324f2df317859ff894db4da9cc4f2cb1ec4fe2641130c89942d11f15" exitCode=137 Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.594082 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerDied","Data":"735e5ebf324f2df317859ff894db4da9cc4f2cb1ec4fe2641130c89942d11f15"} Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.594118 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c","Type":"ContainerDied","Data":"6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb"} Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.594132 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ad8aed316ac4049b66491c040bc7fa86f02386756b23ca16a483363872e91fb" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.617424 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693636 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693770 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693813 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693860 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693897 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvflz\" (UniqueName: \"kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.693975 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.694115 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data\") pod \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\" (UID: \"1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c\") " Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.701342 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.701695 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs" (OuterVolumeSpecName: "logs") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.707371 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.723495 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz" (OuterVolumeSpecName: "kube-api-access-wvflz") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "kube-api-access-wvflz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.730450 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts" (OuterVolumeSpecName: "scripts") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.741915 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.787998 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data" (OuterVolumeSpecName: "config-data") pod "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" (UID: "1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798101 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvflz\" (UniqueName: \"kubernetes.io/projected/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-kube-api-access-wvflz\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798139 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798152 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798164 4751 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798196 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798208 4751 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:06 crc kubenswrapper[4751]: I1002 11:17:06.798219 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.227478 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.227951 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-log" containerID="cri-o://7ed61b6e52b31cf20bff3c00c44e3f575bd256d061571e9f7d5e9ee34f23d14e" gracePeriod=30 Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.228082 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-httpd" containerID="cri-o://11ea07032f163f0087cf6b0727e96c2757f81dd4a7ff8eb23db199c0c6b67ba7" gracePeriod=30 Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.608806 4751 generic.go:334] "Generic (PLEG): container finished" podID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerID="7ed61b6e52b31cf20bff3c00c44e3f575bd256d061571e9f7d5e9ee34f23d14e" exitCode=143 Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.608961 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.609870 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerDied","Data":"7ed61b6e52b31cf20bff3c00c44e3f575bd256d061571e9f7d5e9ee34f23d14e"} Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.647070 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.656666 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.709323 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:17:07 crc kubenswrapper[4751]: E1002 11:17:07.711414 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api-log" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.711440 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api-log" Oct 02 11:17:07 crc kubenswrapper[4751]: E1002 11:17:07.711465 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.711472 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.712065 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.712148 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" containerName="cinder-api-log" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.721131 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.727751 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.728425 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.728680 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.740553 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832374 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9389a471-3750-414f-889f-389fbb54e248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832422 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832456 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832476 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-scripts\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832490 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9389a471-3750-414f-889f-389fbb54e248-logs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832527 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832553 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data-custom\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832594 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.832617 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5nsp\" (UniqueName: \"kubernetes.io/projected/9389a471-3750-414f-889f-389fbb54e248-kube-api-access-m5nsp\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.934593 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9389a471-3750-414f-889f-389fbb54e248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.935314 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.935359 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.934818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9389a471-3750-414f-889f-389fbb54e248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.935387 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-scripts\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.935725 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9389a471-3750-414f-889f-389fbb54e248-logs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.936000 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.936151 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data-custom\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.936481 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.936658 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9389a471-3750-414f-889f-389fbb54e248-logs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.937515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5nsp\" (UniqueName: \"kubernetes.io/projected/9389a471-3750-414f-889f-389fbb54e248-kube-api-access-m5nsp\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.941032 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data-custom\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.941424 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-config-data\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.948273 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.953996 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.954073 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.954413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9389a471-3750-414f-889f-389fbb54e248-scripts\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:07 crc kubenswrapper[4751]: I1002 11:17:07.961155 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5nsp\" (UniqueName: \"kubernetes.io/projected/9389a471-3750-414f-889f-389fbb54e248-kube-api-access-m5nsp\") pod \"cinder-api-0\" (UID: \"9389a471-3750-414f-889f-389fbb54e248\") " pod="openstack/cinder-api-0" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.074450 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.173703 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.243262 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config\") pod \"7c4a8a78-09c8-4928-9425-858e29d498f0\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.243769 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config\") pod \"7c4a8a78-09c8-4928-9425-858e29d498f0\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.243808 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9xrj\" (UniqueName: \"kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj\") pod \"7c4a8a78-09c8-4928-9425-858e29d498f0\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.243937 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle\") pod \"7c4a8a78-09c8-4928-9425-858e29d498f0\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.244227 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs\") pod \"7c4a8a78-09c8-4928-9425-858e29d498f0\" (UID: \"7c4a8a78-09c8-4928-9425-858e29d498f0\") " Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.250945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7c4a8a78-09c8-4928-9425-858e29d498f0" (UID: "7c4a8a78-09c8-4928-9425-858e29d498f0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.253080 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj" (OuterVolumeSpecName: "kube-api-access-b9xrj") pod "7c4a8a78-09c8-4928-9425-858e29d498f0" (UID: "7c4a8a78-09c8-4928-9425-858e29d498f0"). InnerVolumeSpecName "kube-api-access-b9xrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.349755 4751 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.349795 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9xrj\" (UniqueName: \"kubernetes.io/projected/7c4a8a78-09c8-4928-9425-858e29d498f0-kube-api-access-b9xrj\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.350477 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c4a8a78-09c8-4928-9425-858e29d498f0" (UID: "7c4a8a78-09c8-4928-9425-858e29d498f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.366522 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config" (OuterVolumeSpecName: "config") pod "7c4a8a78-09c8-4928-9425-858e29d498f0" (UID: "7c4a8a78-09c8-4928-9425-858e29d498f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.379233 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.379834 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-log" containerID="cri-o://5cb1cdbfff03f8989608a29c15b35b95b8e7743d1f23ff10e9d5835d88718b2d" gracePeriod=30 Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.380157 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-httpd" containerID="cri-o://4f7ff3de9998d2d398821bd2603fc30726046078935698f5fcc731974d93ecc2" gracePeriod=30 Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.400145 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7c4a8a78-09c8-4928-9425-858e29d498f0" (UID: "7c4a8a78-09c8-4928-9425-858e29d498f0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.459502 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.459563 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.459576 4751 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c4a8a78-09c8-4928-9425-858e29d498f0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.644141 4751 generic.go:334] "Generic (PLEG): container finished" podID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerID="288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d" exitCode=0 Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.644748 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9c6c568d-bsnqp" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.648583 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerDied","Data":"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d"} Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.648647 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9c6c568d-bsnqp" event={"ID":"7c4a8a78-09c8-4928-9425-858e29d498f0","Type":"ContainerDied","Data":"3db1ca90e6f43842f5c7034fafea1da20ecc0d8e00126cd0d017622e384373db"} Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.648672 4751 scope.go:117] "RemoveContainer" containerID="3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.677773 4751 generic.go:334] "Generic (PLEG): container finished" podID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerID="5cb1cdbfff03f8989608a29c15b35b95b8e7743d1f23ff10e9d5835d88718b2d" exitCode=143 Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.677847 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerDied","Data":"5cb1cdbfff03f8989608a29c15b35b95b8e7743d1f23ff10e9d5835d88718b2d"} Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.709946 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.717773 4751 scope.go:117] "RemoveContainer" containerID="288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.720828 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.746946 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d9c6c568d-bsnqp"] Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.747437 4751 scope.go:117] "RemoveContainer" containerID="3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9" Oct 02 11:17:08 crc kubenswrapper[4751]: E1002 11:17:08.747802 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9\": container with ID starting with 3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9 not found: ID does not exist" containerID="3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.747832 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9"} err="failed to get container status \"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9\": rpc error: code = NotFound desc = could not find container \"3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9\": container with ID starting with 3b0c66b2c019ab1802b31568dfe48585c8b32db10d64a3cbaf494caa85a8f5e9 not found: ID does not exist" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.747855 4751 scope.go:117] "RemoveContainer" containerID="288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d" Oct 02 11:17:08 crc kubenswrapper[4751]: E1002 11:17:08.748115 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d\": container with ID starting with 288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d not found: ID does not exist" containerID="288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d" Oct 02 11:17:08 crc kubenswrapper[4751]: I1002 11:17:08.748158 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d"} err="failed to get container status \"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d\": rpc error: code = NotFound desc = could not find container \"288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d\": container with ID starting with 288e7922abbe288e3178b0c78819a7e6d42c548142dea10e3cb12f85565e477d not found: ID does not exist" Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.562359 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c" path="/var/lib/kubelet/pods/1d4e3bbb-dc71-4be4-b355-eb1b3db0c70c/volumes" Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.563803 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" path="/var/lib/kubelet/pods/7c4a8a78-09c8-4928-9425-858e29d498f0/volumes" Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.587676 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.588101 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-central-agent" containerID="cri-o://0028e9e60d949d513a2eac9c0e58d7495bf81bb81c42320bef3126557c644663" gracePeriod=30 Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.588145 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="proxy-httpd" containerID="cri-o://d4836bf48fcf8388404521646d5ed11cd88ea2dca5632cf5233ffa9237155d8b" gracePeriod=30 Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.588319 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="sg-core" containerID="cri-o://0dc7b6ba193eb578d87cfd0f949ac57008c192ec0ee6dd2c3f88d28db1065830" gracePeriod=30 Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.588987 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-notification-agent" containerID="cri-o://300f2889f589fdeea428d3f849bf8c0a27c4571ca76789e371de4ae4dd7b06e4" gracePeriod=30 Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.698537 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9389a471-3750-414f-889f-389fbb54e248","Type":"ContainerStarted","Data":"42adc3ecba86c95bada1fb7fd8eb1feecf7349efb44797e37b4306c0cca65af1"} Oct 02 11:17:09 crc kubenswrapper[4751]: I1002 11:17:09.698593 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9389a471-3750-414f-889f-389fbb54e248","Type":"ContainerStarted","Data":"8186d3aa3e0b8ebe449f9b8065a8ce041d17f903e76fccf7fea6ca15990da2c7"} Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.714784 4751 generic.go:334] "Generic (PLEG): container finished" podID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerID="d4836bf48fcf8388404521646d5ed11cd88ea2dca5632cf5233ffa9237155d8b" exitCode=0 Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715133 4751 generic.go:334] "Generic (PLEG): container finished" podID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerID="0dc7b6ba193eb578d87cfd0f949ac57008c192ec0ee6dd2c3f88d28db1065830" exitCode=2 Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715149 4751 generic.go:334] "Generic (PLEG): container finished" podID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerID="300f2889f589fdeea428d3f849bf8c0a27c4571ca76789e371de4ae4dd7b06e4" exitCode=0 Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715163 4751 generic.go:334] "Generic (PLEG): container finished" podID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerID="0028e9e60d949d513a2eac9c0e58d7495bf81bb81c42320bef3126557c644663" exitCode=0 Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.714967 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerDied","Data":"d4836bf48fcf8388404521646d5ed11cd88ea2dca5632cf5233ffa9237155d8b"} Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715253 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerDied","Data":"0dc7b6ba193eb578d87cfd0f949ac57008c192ec0ee6dd2c3f88d28db1065830"} Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715269 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerDied","Data":"300f2889f589fdeea428d3f849bf8c0a27c4571ca76789e371de4ae4dd7b06e4"} Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.715283 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerDied","Data":"0028e9e60d949d513a2eac9c0e58d7495bf81bb81c42320bef3126557c644663"} Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.719429 4751 generic.go:334] "Generic (PLEG): container finished" podID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerID="11ea07032f163f0087cf6b0727e96c2757f81dd4a7ff8eb23db199c0c6b67ba7" exitCode=0 Oct 02 11:17:10 crc kubenswrapper[4751]: I1002 11:17:10.719461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerDied","Data":"11ea07032f163f0087cf6b0727e96c2757f81dd4a7ff8eb23db199c0c6b67ba7"} Oct 02 11:17:11 crc kubenswrapper[4751]: I1002 11:17:11.731514 4751 generic.go:334] "Generic (PLEG): container finished" podID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerID="4f7ff3de9998d2d398821bd2603fc30726046078935698f5fcc731974d93ecc2" exitCode=0 Oct 02 11:17:11 crc kubenswrapper[4751]: I1002 11:17:11.731561 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerDied","Data":"4f7ff3de9998d2d398821bd2603fc30726046078935698f5fcc731974d93ecc2"} Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.523071 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:12 crc kubenswrapper[4751]: E1002 11:17:12.523927 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-httpd" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.523950 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-httpd" Oct 02 11:17:12 crc kubenswrapper[4751]: E1002 11:17:12.523987 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-api" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.523999 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-api" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.524334 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-httpd" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.524362 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4a8a78-09c8-4928-9425-858e29d498f0" containerName="neutron-api" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.528371 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.539722 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.650919 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.651257 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ccrc\" (UniqueName: \"kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.651297 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.752741 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.752832 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ccrc\" (UniqueName: \"kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.752855 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.753380 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.753672 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.773229 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ccrc\" (UniqueName: \"kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc\") pod \"redhat-operators-s4c29\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:12 crc kubenswrapper[4751]: I1002 11:17:12.859013 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.289313 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.349811 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402230 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402291 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402319 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402358 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmdp8\" (UniqueName: \"kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402382 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402420 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402477 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402497 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg7rs\" (UniqueName: \"kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402526 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402549 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402642 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402664 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data\") pod \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\" (UID: \"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402684 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402713 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.402740 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data\") pod \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\" (UID: \"f016b483-dbd6-466a-8dd5-ab1a966dfd61\") " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.404676 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs" (OuterVolumeSpecName: "logs") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.407855 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.408247 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.408291 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.416881 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8" (OuterVolumeSpecName: "kube-api-access-fmdp8") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "kube-api-access-fmdp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.417034 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts" (OuterVolumeSpecName: "scripts") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.417119 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.417140 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs" (OuterVolumeSpecName: "kube-api-access-zg7rs") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "kube-api-access-zg7rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.419368 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts" (OuterVolumeSpecName: "scripts") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.441119 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.481946 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505019 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505404 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg7rs\" (UniqueName: \"kubernetes.io/projected/f016b483-dbd6-466a-8dd5-ab1a966dfd61-kube-api-access-zg7rs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505421 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505433 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505444 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505455 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505465 4751 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f016b483-dbd6-466a-8dd5-ab1a966dfd61-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505476 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505509 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505521 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.505532 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmdp8\" (UniqueName: \"kubernetes.io/projected/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-kube-api-access-fmdp8\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.514857 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.532018 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.556067 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.573801 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.579324 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data" (OuterVolumeSpecName: "config-data") pod "f016b483-dbd6-466a-8dd5-ab1a966dfd61" (UID: "f016b483-dbd6-466a-8dd5-ab1a966dfd61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.607792 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.607832 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.607842 4751 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f016b483-dbd6-466a-8dd5-ab1a966dfd61-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.607851 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.619581 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data" (OuterVolumeSpecName: "config-data") pod "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" (UID: "7327d94a-eef2-4d01-b9e3-dca4d5fb70c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.709733 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.781355 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" event={"ID":"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d","Type":"ContainerStarted","Data":"182c3a4c0c5995c735c441f652a5dd7e40f34bb89cd11b97ac14e1d6cac2edde"} Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.785048 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.785156 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7327d94a-eef2-4d01-b9e3-dca4d5fb70c4","Type":"ContainerDied","Data":"0c914d790c9d0174f73923a2091d0ada0bde3847ccc0a9a4003e27034860eb15"} Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.785217 4751 scope.go:117] "RemoveContainer" containerID="d4836bf48fcf8388404521646d5ed11cd88ea2dca5632cf5233ffa9237155d8b" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.794510 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f016b483-dbd6-466a-8dd5-ab1a966dfd61","Type":"ContainerDied","Data":"1be9726ec684bf1608b543fe0251e00b8de0c7ebfac8cc33e2442aadd9332ed2"} Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.794580 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.800009 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerStarted","Data":"6797a33b1c82d7f5f87f2b699858f38804c23e799694264a943a5a7cb569af70"} Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.815631 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" podStartSLOduration=1.571240838 podStartE2EDuration="11.815604301s" podCreationTimestamp="2025-10-02 11:17:04 +0000 UTC" firstStartedPulling="2025-10-02 11:17:04.937353553 +0000 UTC m=+1506.991580003" lastFinishedPulling="2025-10-02 11:17:15.181716986 +0000 UTC m=+1517.235943466" observedRunningTime="2025-10-02 11:17:15.796818279 +0000 UTC m=+1517.851044759" watchObservedRunningTime="2025-10-02 11:17:15.815604301 +0000 UTC m=+1517.869830771" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.837744 4751 scope.go:117] "RemoveContainer" containerID="0dc7b6ba193eb578d87cfd0f949ac57008c192ec0ee6dd2c3f88d28db1065830" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.845402 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.865895 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.883078 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.904486 4751 scope.go:117] "RemoveContainer" containerID="300f2889f589fdeea428d3f849bf8c0a27c4571ca76789e371de4ae4dd7b06e4" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910307 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910731 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-notification-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910748 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-notification-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910767 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910774 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910792 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-log" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910800 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-log" Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910818 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="proxy-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910825 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="proxy-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910841 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-central-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910848 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-central-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: E1002 11:17:15.910858 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="sg-core" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.910864 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="sg-core" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911033 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911054 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" containerName="glance-log" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911065 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-notification-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911076 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="proxy-httpd" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911083 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="sg-core" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.911094 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" containerName="ceilometer-central-agent" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.916714 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.919623 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.919813 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.921852 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.940666 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.943451 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.946275 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.946386 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.949481 4751 scope.go:117] "RemoveContainer" containerID="0028e9e60d949d513a2eac9c0e58d7495bf81bb81c42320bef3126557c644663" Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.950861 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.969946 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:15 crc kubenswrapper[4751]: I1002 11:17:15.973780 4751 scope.go:117] "RemoveContainer" containerID="11ea07032f163f0087cf6b0727e96c2757f81dd4a7ff8eb23db199c0c6b67ba7" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.002235 4751 scope.go:117] "RemoveContainer" containerID="7ed61b6e52b31cf20bff3c00c44e3f575bd256d061571e9f7d5e9ee34f23d14e" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016206 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016253 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016285 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016300 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016326 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016348 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-logs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016374 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sd4s\" (UniqueName: \"kubernetes.io/projected/f1a17d69-9670-4f40-89dd-01099655729a-kube-api-access-2sd4s\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016411 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016470 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016494 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjxt8\" (UniqueName: \"kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016517 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016539 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016564 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016585 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.016602 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118412 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118469 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118512 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118547 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-logs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118588 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sd4s\" (UniqueName: \"kubernetes.io/projected/f1a17d69-9670-4f40-89dd-01099655729a-kube-api-access-2sd4s\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118607 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118634 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118658 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjxt8\" (UniqueName: \"kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118681 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118712 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118743 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118769 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118793 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118880 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.118902 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.119117 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.119216 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1a17d69-9670-4f40-89dd-01099655729a-logs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.119666 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.119986 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.123880 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.125212 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.125417 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-config-data\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.125865 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.126416 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.127082 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.127366 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-scripts\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.129635 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.129927 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1a17d69-9670-4f40-89dd-01099655729a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.138072 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sd4s\" (UniqueName: \"kubernetes.io/projected/f1a17d69-9670-4f40-89dd-01099655729a-kube-api-access-2sd4s\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.141103 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjxt8\" (UniqueName: \"kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8\") pod \"ceilometer-0\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.153959 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f1a17d69-9670-4f40-89dd-01099655729a\") " pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.239478 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.270010 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.424997 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524244 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524412 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524640 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524695 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524778 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524869 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5brxm\" (UniqueName: \"kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524936 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.524963 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs\") pod \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\" (UID: \"d775ae5c-13ff-434f-be74-bb9bd70a0e4e\") " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.526673 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs" (OuterVolumeSpecName: "logs") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.527885 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.539010 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.542657 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm" (OuterVolumeSpecName: "kube-api-access-5brxm") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "kube-api-access-5brxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.548458 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts" (OuterVolumeSpecName: "scripts") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.577992 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.607554 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.627790 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data" (OuterVolumeSpecName: "config-data") pod "d775ae5c-13ff-434f-be74-bb9bd70a0e4e" (UID: "d775ae5c-13ff-434f-be74-bb9bd70a0e4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628013 4751 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628035 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628048 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628064 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5brxm\" (UniqueName: \"kubernetes.io/projected/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-kube-api-access-5brxm\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628078 4751 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628090 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628132 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.628357 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d775ae5c-13ff-434f-be74-bb9bd70a0e4e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.687389 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 11:17:16 crc kubenswrapper[4751]: E1002 11:17:16.726938 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.740428 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.744601 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.821894 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerStarted","Data":"354e77288956bdfc70f959f02349733a9d23a4afc60601af761e52483a79c55a"} Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.825691 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d775ae5c-13ff-434f-be74-bb9bd70a0e4e","Type":"ContainerDied","Data":"c987f12462ec5e8259dee7a0b8437776673588280bb6ba3ffa1a9bbf4f7cac85"} Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.825735 4751 scope.go:117] "RemoveContainer" containerID="4f7ff3de9998d2d398821bd2603fc30726046078935698f5fcc731974d93ecc2" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.825787 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.835489 4751 generic.go:334] "Generic (PLEG): container finished" podID="f6220425-70aa-4230-93b3-074467ddaa75" containerID="ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e" exitCode=0 Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.835554 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerDied","Data":"ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e"} Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.848641 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9389a471-3750-414f-889f-389fbb54e248","Type":"ContainerStarted","Data":"684d721da6bc1307f32136ff9c953cb36907f77a94c3aecd06d68d7949f46662"} Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.848914 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.880427 4751 scope.go:117] "RemoveContainer" containerID="5cb1cdbfff03f8989608a29c15b35b95b8e7743d1f23ff10e9d5835d88718b2d" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.924070 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.933957 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.953240 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:16 crc kubenswrapper[4751]: E1002 11:17:16.953664 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-log" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.953678 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-log" Oct 02 11:17:16 crc kubenswrapper[4751]: E1002 11:17:16.953707 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-httpd" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.953713 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-httpd" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.953891 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-httpd" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.953921 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" containerName="glance-log" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.954872 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.957730 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.957934 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.958526 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=9.958499443000001 podStartE2EDuration="9.958499443s" podCreationTimestamp="2025-10-02 11:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:16.903574684 +0000 UTC m=+1518.957801134" watchObservedRunningTime="2025-10-02 11:17:16.958499443 +0000 UTC m=+1519.012725893" Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.973674 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:16 crc kubenswrapper[4751]: I1002 11:17:16.998664 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 11:17:17 crc kubenswrapper[4751]: W1002 11:17:17.006137 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1a17d69_9670_4f40_89dd_01099655729a.slice/crio-8eb80cfbbd62196b483c3c4f91415f3046304fb2b86280848f67c0ed170c9bca WatchSource:0}: Error finding container 8eb80cfbbd62196b483c3c4f91415f3046304fb2b86280848f67c0ed170c9bca: Status 404 returned error can't find the container with id 8eb80cfbbd62196b483c3c4f91415f3046304fb2b86280848f67c0ed170c9bca Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055562 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055694 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055736 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055772 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpknb\" (UniqueName: \"kubernetes.io/projected/4ded07b6-69f6-4fed-9993-5877f61a4bf8-kube-api-access-fpknb\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055826 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055904 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055956 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.055980 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157407 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpknb\" (UniqueName: \"kubernetes.io/projected/4ded07b6-69f6-4fed-9993-5877f61a4bf8-kube-api-access-fpknb\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157459 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157510 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157542 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157556 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157601 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157816 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.157837 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.158202 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.158522 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-logs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.163930 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ded07b6-69f6-4fed-9993-5877f61a4bf8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.169548 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.170594 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.172063 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.172327 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ded07b6-69f6-4fed-9993-5877f61a4bf8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.178714 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpknb\" (UniqueName: \"kubernetes.io/projected/4ded07b6-69f6-4fed-9993-5877f61a4bf8-kube-api-access-fpknb\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.190575 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"4ded07b6-69f6-4fed-9993-5877f61a4bf8\") " pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.288593 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.569329 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7327d94a-eef2-4d01-b9e3-dca4d5fb70c4" path="/var/lib/kubelet/pods/7327d94a-eef2-4d01-b9e3-dca4d5fb70c4/volumes" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.570492 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d775ae5c-13ff-434f-be74-bb9bd70a0e4e" path="/var/lib/kubelet/pods/d775ae5c-13ff-434f-be74-bb9bd70a0e4e/volumes" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.571911 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f016b483-dbd6-466a-8dd5-ab1a966dfd61" path="/var/lib/kubelet/pods/f016b483-dbd6-466a-8dd5-ab1a966dfd61/volumes" Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.871935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerStarted","Data":"efa444739f95ded125966a8e41ecf8abbe44868f49908672a62d78b1fdeea688"} Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.872439 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.875909 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1a17d69-9670-4f40-89dd-01099655729a","Type":"ContainerStarted","Data":"a35f08eea4414c204d1c7dc2d1b9a76e7471106a7048e14630e79e9d26ed83a8"} Oct 02 11:17:17 crc kubenswrapper[4751]: I1002 11:17:17.875942 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1a17d69-9670-4f40-89dd-01099655729a","Type":"ContainerStarted","Data":"8eb80cfbbd62196b483c3c4f91415f3046304fb2b86280848f67c0ed170c9bca"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.891356 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerStarted","Data":"54c487812acd03ecde14f6528ebf21dce4be46c91cc5b57bd8938a831d09711c"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.906685 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f1a17d69-9670-4f40-89dd-01099655729a","Type":"ContainerStarted","Data":"b2a10c2f05b0b235ed22392227de909efadb32c1a06fd469c112285cc9334d64"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.917108 4751 generic.go:334] "Generic (PLEG): container finished" podID="f6220425-70aa-4230-93b3-074467ddaa75" containerID="628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c" exitCode=0 Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.917200 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerDied","Data":"628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.943628 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ded07b6-69f6-4fed-9993-5877f61a4bf8","Type":"ContainerStarted","Data":"cd38af7353912eaa2ec4ef42361c747a4c1ad87e3181511e95cd7212769572e9"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.943676 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ded07b6-69f6-4fed-9993-5877f61a4bf8","Type":"ContainerStarted","Data":"410599a1f48cf32d4d2c51e232439bd1483669ea5c6c25fe361818430c9f6a40"} Oct 02 11:17:18 crc kubenswrapper[4751]: I1002 11:17:18.994639 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.994620491 podStartE2EDuration="3.994620491s" podCreationTimestamp="2025-10-02 11:17:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:18.944116901 +0000 UTC m=+1520.998343351" watchObservedRunningTime="2025-10-02 11:17:18.994620491 +0000 UTC m=+1521.048846941" Oct 02 11:17:19 crc kubenswrapper[4751]: I1002 11:17:19.956762 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerStarted","Data":"e3943a58580606ecbf7a04c6a558b6adf8c6251d88865d401f322accd4a58722"} Oct 02 11:17:19 crc kubenswrapper[4751]: I1002 11:17:19.960064 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerStarted","Data":"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3"} Oct 02 11:17:19 crc kubenswrapper[4751]: I1002 11:17:19.968457 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4ded07b6-69f6-4fed-9993-5877f61a4bf8","Type":"ContainerStarted","Data":"1959e8ab3a8b02072201042b388c53e0654ab3da05a06828795deb0cf2328b53"} Oct 02 11:17:19 crc kubenswrapper[4751]: I1002 11:17:19.982621 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s4c29" podStartSLOduration=5.479347066 podStartE2EDuration="7.982601202s" podCreationTimestamp="2025-10-02 11:17:12 +0000 UTC" firstStartedPulling="2025-10-02 11:17:16.846087498 +0000 UTC m=+1518.900313948" lastFinishedPulling="2025-10-02 11:17:19.349341634 +0000 UTC m=+1521.403568084" observedRunningTime="2025-10-02 11:17:19.977804204 +0000 UTC m=+1522.032030654" watchObservedRunningTime="2025-10-02 11:17:19.982601202 +0000 UTC m=+1522.036827672" Oct 02 11:17:20 crc kubenswrapper[4751]: I1002 11:17:20.000625 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.000602443 podStartE2EDuration="4.000602443s" podCreationTimestamp="2025-10-02 11:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:19.999446843 +0000 UTC m=+1522.053673283" watchObservedRunningTime="2025-10-02 11:17:20.000602443 +0000 UTC m=+1522.054828903" Oct 02 11:17:20 crc kubenswrapper[4751]: I1002 11:17:20.977348 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerStarted","Data":"770b8ba699fda9d0f0ce88975e8848a4271b6081881becca62ecd48af60a5c47"} Oct 02 11:17:20 crc kubenswrapper[4751]: I1002 11:17:20.977980 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:17:21 crc kubenswrapper[4751]: I1002 11:17:21.024838 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.632307584 podStartE2EDuration="6.024814632s" podCreationTimestamp="2025-10-02 11:17:15 +0000 UTC" firstStartedPulling="2025-10-02 11:17:16.752965148 +0000 UTC m=+1518.807191598" lastFinishedPulling="2025-10-02 11:17:20.145472196 +0000 UTC m=+1522.199698646" observedRunningTime="2025-10-02 11:17:21.018821232 +0000 UTC m=+1523.073047692" watchObservedRunningTime="2025-10-02 11:17:21.024814632 +0000 UTC m=+1523.079041102" Oct 02 11:17:22 crc kubenswrapper[4751]: I1002 11:17:22.861564 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:22 crc kubenswrapper[4751]: I1002 11:17:22.861818 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:23 crc kubenswrapper[4751]: I1002 11:17:23.908019 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s4c29" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="registry-server" probeResult="failure" output=< Oct 02 11:17:23 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 11:17:23 crc kubenswrapper[4751]: > Oct 02 11:17:26 crc kubenswrapper[4751]: I1002 11:17:26.008569 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 11:17:26 crc kubenswrapper[4751]: I1002 11:17:26.270357 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 11:17:26 crc kubenswrapper[4751]: I1002 11:17:26.270683 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 11:17:26 crc kubenswrapper[4751]: I1002 11:17:26.308894 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 11:17:26 crc kubenswrapper[4751]: I1002 11:17:26.321677 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 11:17:26 crc kubenswrapper[4751]: E1002 11:17:26.986011 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.031345 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.031442 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.289392 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.289755 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.318626 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:27 crc kubenswrapper[4751]: I1002 11:17:27.327121 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:28 crc kubenswrapper[4751]: I1002 11:17:28.039966 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:28 crc kubenswrapper[4751]: I1002 11:17:28.041650 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:29 crc kubenswrapper[4751]: I1002 11:17:29.850369 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 11:17:29 crc kubenswrapper[4751]: I1002 11:17:29.852035 4751 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 11:17:29 crc kubenswrapper[4751]: I1002 11:17:29.886538 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 11:17:29 crc kubenswrapper[4751]: I1002 11:17:29.943385 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:29 crc kubenswrapper[4751]: I1002 11:17:29.959329 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 11:17:32 crc kubenswrapper[4751]: I1002 11:17:32.926267 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:32 crc kubenswrapper[4751]: I1002 11:17:32.976934 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:33 crc kubenswrapper[4751]: I1002 11:17:33.161601 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.102108 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s4c29" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="registry-server" containerID="cri-o://33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3" gracePeriod=2 Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.595949 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.723848 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities\") pod \"f6220425-70aa-4230-93b3-074467ddaa75\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.724116 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ccrc\" (UniqueName: \"kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc\") pod \"f6220425-70aa-4230-93b3-074467ddaa75\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.724197 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content\") pod \"f6220425-70aa-4230-93b3-074467ddaa75\" (UID: \"f6220425-70aa-4230-93b3-074467ddaa75\") " Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.725410 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities" (OuterVolumeSpecName: "utilities") pod "f6220425-70aa-4230-93b3-074467ddaa75" (UID: "f6220425-70aa-4230-93b3-074467ddaa75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.730753 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc" (OuterVolumeSpecName: "kube-api-access-8ccrc") pod "f6220425-70aa-4230-93b3-074467ddaa75" (UID: "f6220425-70aa-4230-93b3-074467ddaa75"). InnerVolumeSpecName "kube-api-access-8ccrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.826872 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.826910 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ccrc\" (UniqueName: \"kubernetes.io/projected/f6220425-70aa-4230-93b3-074467ddaa75-kube-api-access-8ccrc\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.835195 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6220425-70aa-4230-93b3-074467ddaa75" (UID: "f6220425-70aa-4230-93b3-074467ddaa75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:34 crc kubenswrapper[4751]: I1002 11:17:34.928391 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6220425-70aa-4230-93b3-074467ddaa75-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.115908 4751 generic.go:334] "Generic (PLEG): container finished" podID="f6220425-70aa-4230-93b3-074467ddaa75" containerID="33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3" exitCode=0 Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.115997 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerDied","Data":"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3"} Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.116040 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4c29" event={"ID":"f6220425-70aa-4230-93b3-074467ddaa75","Type":"ContainerDied","Data":"6797a33b1c82d7f5f87f2b699858f38804c23e799694264a943a5a7cb569af70"} Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.116076 4751 scope.go:117] "RemoveContainer" containerID="33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.116094 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4c29" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.147122 4751 scope.go:117] "RemoveContainer" containerID="628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.162388 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.173283 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s4c29"] Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.181910 4751 scope.go:117] "RemoveContainer" containerID="ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.220779 4751 scope.go:117] "RemoveContainer" containerID="33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3" Oct 02 11:17:35 crc kubenswrapper[4751]: E1002 11:17:35.221223 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3\": container with ID starting with 33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3 not found: ID does not exist" containerID="33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.221268 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3"} err="failed to get container status \"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3\": rpc error: code = NotFound desc = could not find container \"33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3\": container with ID starting with 33a65bfc7d8a90c966c7ffddd192df768810b156b93e5957e8132aa19973a5c3 not found: ID does not exist" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.221297 4751 scope.go:117] "RemoveContainer" containerID="628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c" Oct 02 11:17:35 crc kubenswrapper[4751]: E1002 11:17:35.221686 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c\": container with ID starting with 628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c not found: ID does not exist" containerID="628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.221722 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c"} err="failed to get container status \"628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c\": rpc error: code = NotFound desc = could not find container \"628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c\": container with ID starting with 628dde150acc432fba52215ab262d9236ad474a80e616ed9b2c6117bc33f392c not found: ID does not exist" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.221739 4751 scope.go:117] "RemoveContainer" containerID="ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e" Oct 02 11:17:35 crc kubenswrapper[4751]: E1002 11:17:35.222101 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e\": container with ID starting with ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e not found: ID does not exist" containerID="ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.222130 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e"} err="failed to get container status \"ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e\": rpc error: code = NotFound desc = could not find container \"ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e\": container with ID starting with ea8a6f97fb14a64709fa482ae5a9b460114f2ff3e00236cd812668c19dedd86e not found: ID does not exist" Oct 02 11:17:35 crc kubenswrapper[4751]: I1002 11:17:35.561301 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6220425-70aa-4230-93b3-074467ddaa75" path="/var/lib/kubelet/pods/f6220425-70aa-4230-93b3-074467ddaa75/volumes" Oct 02 11:17:37 crc kubenswrapper[4751]: E1002 11:17:37.219606 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:39 crc kubenswrapper[4751]: I1002 11:17:39.162078 4751 generic.go:334] "Generic (PLEG): container finished" podID="a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" containerID="182c3a4c0c5995c735c441f652a5dd7e40f34bb89cd11b97ac14e1d6cac2edde" exitCode=0 Oct 02 11:17:39 crc kubenswrapper[4751]: I1002 11:17:39.162146 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" event={"ID":"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d","Type":"ContainerDied","Data":"182c3a4c0c5995c735c441f652a5dd7e40f34bb89cd11b97ac14e1d6cac2edde"} Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.532288 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.630349 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g6xl\" (UniqueName: \"kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl\") pod \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.630474 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data\") pod \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.630502 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle\") pod \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.630708 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts\") pod \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\" (UID: \"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d\") " Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.639690 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts" (OuterVolumeSpecName: "scripts") pod "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" (UID: "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.639719 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl" (OuterVolumeSpecName: "kube-api-access-8g6xl") pod "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" (UID: "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d"). InnerVolumeSpecName "kube-api-access-8g6xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.663333 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data" (OuterVolumeSpecName: "config-data") pod "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" (UID: "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.678116 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" (UID: "a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.733642 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.733696 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g6xl\" (UniqueName: \"kubernetes.io/projected/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-kube-api-access-8g6xl\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.733718 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:40 crc kubenswrapper[4751]: I1002 11:17:40.733737 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.184496 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" event={"ID":"a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d","Type":"ContainerDied","Data":"583ee1d11b11fd9d350c24ea25f1c758ca320d11f7fbbbb8fc3a4c46594f5aa8"} Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.185113 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="583ee1d11b11fd9d350c24ea25f1c758ca320d11f7fbbbb8fc3a4c46594f5aa8" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.184579 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-q7dsc" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.307634 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 11:17:41 crc kubenswrapper[4751]: E1002 11:17:41.307996 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" containerName="nova-cell0-conductor-db-sync" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308015 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" containerName="nova-cell0-conductor-db-sync" Oct 02 11:17:41 crc kubenswrapper[4751]: E1002 11:17:41.308036 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="registry-server" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308044 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="registry-server" Oct 02 11:17:41 crc kubenswrapper[4751]: E1002 11:17:41.308070 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="extract-content" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308076 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="extract-content" Oct 02 11:17:41 crc kubenswrapper[4751]: E1002 11:17:41.308088 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="extract-utilities" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308094 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="extract-utilities" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308306 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6220425-70aa-4230-93b3-074467ddaa75" containerName="registry-server" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.308328 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" containerName="nova-cell0-conductor-db-sync" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.309090 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.316519 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.321129 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kjgf2" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.324699 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.444407 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.444922 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.445039 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpsf4\" (UniqueName: \"kubernetes.io/projected/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-kube-api-access-rpsf4\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.546491 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.546572 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.546604 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpsf4\" (UniqueName: \"kubernetes.io/projected/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-kube-api-access-rpsf4\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.552053 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.552451 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.575879 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpsf4\" (UniqueName: \"kubernetes.io/projected/aa5e8e06-7063-4bb7-85dc-fe696c75aa7e-kube-api-access-rpsf4\") pod \"nova-cell0-conductor-0\" (UID: \"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:41 crc kubenswrapper[4751]: I1002 11:17:41.628202 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.069857 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.193241 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.195783 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.212423 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.216053 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e","Type":"ContainerStarted","Data":"0b626c998c40187bcfb02e9f4aaddc0469bb4d8ecbf43f9d835779d5c7c80643"} Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.264835 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.264909 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx4w6\" (UniqueName: \"kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.264998 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.366417 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.366503 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx4w6\" (UniqueName: \"kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.366578 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.367023 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.367140 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.385568 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx4w6\" (UniqueName: \"kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6\") pod \"certified-operators-5qfmq\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:42 crc kubenswrapper[4751]: I1002 11:17:42.532019 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.014847 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:43 crc kubenswrapper[4751]: W1002 11:17:43.017543 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e0ef4bb_bff2_4d47_bc75_0af7a4463fc0.slice/crio-eb5b83f5bfe8bc78b40686058bd86ea0ee314415dcbef306e6813c1c9409eef3 WatchSource:0}: Error finding container eb5b83f5bfe8bc78b40686058bd86ea0ee314415dcbef306e6813c1c9409eef3: Status 404 returned error can't find the container with id eb5b83f5bfe8bc78b40686058bd86ea0ee314415dcbef306e6813c1c9409eef3 Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.226813 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"aa5e8e06-7063-4bb7-85dc-fe696c75aa7e","Type":"ContainerStarted","Data":"835e9cfd960eadba4eb732ce9ee5319b9707df960c0b1833bb03afaeaaf93ae5"} Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.227475 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.229488 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerStarted","Data":"9e40b94370dad0395be96ec3a364aee3f70e9f8bd89234b4cd25df98ca5417df"} Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.229522 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerStarted","Data":"eb5b83f5bfe8bc78b40686058bd86ea0ee314415dcbef306e6813c1c9409eef3"} Oct 02 11:17:43 crc kubenswrapper[4751]: I1002 11:17:43.255092 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.255074077 podStartE2EDuration="2.255074077s" podCreationTimestamp="2025-10-02 11:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:43.248043659 +0000 UTC m=+1545.302270109" watchObservedRunningTime="2025-10-02 11:17:43.255074077 +0000 UTC m=+1545.309300527" Oct 02 11:17:44 crc kubenswrapper[4751]: I1002 11:17:44.241358 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerID="9e40b94370dad0395be96ec3a364aee3f70e9f8bd89234b4cd25df98ca5417df" exitCode=0 Oct 02 11:17:44 crc kubenswrapper[4751]: I1002 11:17:44.243010 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerDied","Data":"9e40b94370dad0395be96ec3a364aee3f70e9f8bd89234b4cd25df98ca5417df"} Oct 02 11:17:46 crc kubenswrapper[4751]: I1002 11:17:46.247936 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 11:17:46 crc kubenswrapper[4751]: I1002 11:17:46.262695 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerID="7e81669d0112becb5f33491d84d37db7fae724ac04c662ba5d497a4746fdcce5" exitCode=0 Oct 02 11:17:46 crc kubenswrapper[4751]: I1002 11:17:46.262777 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerDied","Data":"7e81669d0112becb5f33491d84d37db7fae724ac04c662ba5d497a4746fdcce5"} Oct 02 11:17:47 crc kubenswrapper[4751]: I1002 11:17:47.278966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerStarted","Data":"afd4b0956824898b295f5afee0cf8fde826c093f1d97c943c33b32874b110233"} Oct 02 11:17:47 crc kubenswrapper[4751]: I1002 11:17:47.306762 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5qfmq" podStartSLOduration=2.7575650510000003 podStartE2EDuration="5.306740845s" podCreationTimestamp="2025-10-02 11:17:42 +0000 UTC" firstStartedPulling="2025-10-02 11:17:44.244832525 +0000 UTC m=+1546.299058975" lastFinishedPulling="2025-10-02 11:17:46.794008319 +0000 UTC m=+1548.848234769" observedRunningTime="2025-10-02 11:17:47.304521036 +0000 UTC m=+1549.358747496" watchObservedRunningTime="2025-10-02 11:17:47.306740845 +0000 UTC m=+1549.360967295" Oct 02 11:17:47 crc kubenswrapper[4751]: E1002 11:17:47.478851 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:49 crc kubenswrapper[4751]: I1002 11:17:49.849819 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:49 crc kubenswrapper[4751]: I1002 11:17:49.850934 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" containerName="kube-state-metrics" containerID="cri-o://68a263fe7a00e1b5527ce62c4038d2a1285a89c7b1c52046ab5cf03d8c97af0b" gracePeriod=30 Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.307452 4751 generic.go:334] "Generic (PLEG): container finished" podID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" containerID="68a263fe7a00e1b5527ce62c4038d2a1285a89c7b1c52046ab5cf03d8c97af0b" exitCode=2 Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.307527 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82876497-5a0e-4912-8e2f-e5e5a43f5c95","Type":"ContainerDied","Data":"68a263fe7a00e1b5527ce62c4038d2a1285a89c7b1c52046ab5cf03d8c97af0b"} Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.307776 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"82876497-5a0e-4912-8e2f-e5e5a43f5c95","Type":"ContainerDied","Data":"b3000dd8e26dadbb4f0f0ef9d870effd8690755ae435ee3c3f94414e7e7437b2"} Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.307795 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3000dd8e26dadbb4f0f0ef9d870effd8690755ae435ee3c3f94414e7e7437b2" Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.350328 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.421548 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kghp4\" (UniqueName: \"kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4\") pod \"82876497-5a0e-4912-8e2f-e5e5a43f5c95\" (UID: \"82876497-5a0e-4912-8e2f-e5e5a43f5c95\") " Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.428212 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4" (OuterVolumeSpecName: "kube-api-access-kghp4") pod "82876497-5a0e-4912-8e2f-e5e5a43f5c95" (UID: "82876497-5a0e-4912-8e2f-e5e5a43f5c95"). InnerVolumeSpecName "kube-api-access-kghp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:50 crc kubenswrapper[4751]: I1002 11:17:50.523090 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kghp4\" (UniqueName: \"kubernetes.io/projected/82876497-5a0e-4912-8e2f-e5e5a43f5c95-kube-api-access-kghp4\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.318708 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.355629 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.364601 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.381920 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:51 crc kubenswrapper[4751]: E1002 11:17:51.382663 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" containerName="kube-state-metrics" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.382759 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" containerName="kube-state-metrics" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.382987 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" containerName="kube-state-metrics" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.383706 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.388013 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.393459 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.398401 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.438211 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.438255 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.438334 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.438357 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pptxq\" (UniqueName: \"kubernetes.io/projected/df30039f-e3d2-43ce-aab5-050e2881c224-kube-api-access-pptxq\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.540305 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.540351 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.540409 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.540427 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pptxq\" (UniqueName: \"kubernetes.io/projected/df30039f-e3d2-43ce-aab5-050e2881c224-kube-api-access-pptxq\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.546865 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.546916 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.547965 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df30039f-e3d2-43ce-aab5-050e2881c224-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.558164 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pptxq\" (UniqueName: \"kubernetes.io/projected/df30039f-e3d2-43ce-aab5-050e2881c224-kube-api-access-pptxq\") pod \"kube-state-metrics-0\" (UID: \"df30039f-e3d2-43ce-aab5-050e2881c224\") " pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.562557 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82876497-5a0e-4912-8e2f-e5e5a43f5c95" path="/var/lib/kubelet/pods/82876497-5a0e-4912-8e2f-e5e5a43f5c95/volumes" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.657289 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.720292 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.730353 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.730621 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-central-agent" containerID="cri-o://efa444739f95ded125966a8e41ecf8abbe44868f49908672a62d78b1fdeea688" gracePeriod=30 Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.730701 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="sg-core" containerID="cri-o://e3943a58580606ecbf7a04c6a558b6adf8c6251d88865d401f322accd4a58722" gracePeriod=30 Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.730727 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="proxy-httpd" containerID="cri-o://770b8ba699fda9d0f0ce88975e8848a4271b6081881becca62ecd48af60a5c47" gracePeriod=30 Oct 02 11:17:51 crc kubenswrapper[4751]: I1002 11:17:51.730736 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-notification-agent" containerID="cri-o://54c487812acd03ecde14f6528ebf21dce4be46c91cc5b57bd8938a831d09711c" gracePeriod=30 Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.123344 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-sq6r8"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.124771 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.127193 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.132334 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.143110 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-sq6r8"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.175654 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.175723 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.175797 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.175927 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xgrf\" (UniqueName: \"kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.223500 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: W1002 11:17:52.235898 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf30039f_e3d2_43ce_aab5_050e2881c224.slice/crio-e23d9fcd1d0b5218e515b3e5fd90628f7bbeb83185e142c0cd364e7a9cc5b409 WatchSource:0}: Error finding container e23d9fcd1d0b5218e515b3e5fd90628f7bbeb83185e142c0cd364e7a9cc5b409: Status 404 returned error can't find the container with id e23d9fcd1d0b5218e515b3e5fd90628f7bbeb83185e142c0cd364e7a9cc5b409 Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.282274 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.282615 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xgrf\" (UniqueName: \"kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.282783 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.282863 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.301834 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.308890 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.311524 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xgrf\" (UniqueName: \"kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.314931 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data\") pod \"nova-cell0-cell-mapping-sq6r8\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.339862 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"df30039f-e3d2-43ce-aab5-050e2881c224","Type":"ContainerStarted","Data":"e23d9fcd1d0b5218e515b3e5fd90628f7bbeb83185e142c0cd364e7a9cc5b409"} Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349067 4751 generic.go:334] "Generic (PLEG): container finished" podID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerID="770b8ba699fda9d0f0ce88975e8848a4271b6081881becca62ecd48af60a5c47" exitCode=0 Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349101 4751 generic.go:334] "Generic (PLEG): container finished" podID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerID="e3943a58580606ecbf7a04c6a558b6adf8c6251d88865d401f322accd4a58722" exitCode=2 Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349109 4751 generic.go:334] "Generic (PLEG): container finished" podID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerID="efa444739f95ded125966a8e41ecf8abbe44868f49908672a62d78b1fdeea688" exitCode=0 Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349127 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerDied","Data":"770b8ba699fda9d0f0ce88975e8848a4271b6081881becca62ecd48af60a5c47"} Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349154 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerDied","Data":"e3943a58580606ecbf7a04c6a558b6adf8c6251d88865d401f322accd4a58722"} Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.349201 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerDied","Data":"efa444739f95ded125966a8e41ecf8abbe44868f49908672a62d78b1fdeea688"} Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.353981 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.356135 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.359968 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.367090 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.377420 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.384650 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.385259 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.397958 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.447075 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.477373 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.478524 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.481070 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.495801 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.495902 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.495926 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.495953 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.496052 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s7rg\" (UniqueName: \"kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.496112 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.496155 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54d47\" (UniqueName: \"kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.496223 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.540250 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.540383 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.541990 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.566245 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.567783 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598183 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598238 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gzgh\" (UniqueName: \"kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598266 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54d47\" (UniqueName: \"kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598291 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598331 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598395 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598415 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598433 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598447 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598522 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.598544 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s7rg\" (UniqueName: \"kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.600345 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.601095 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.614518 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.616318 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.616862 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.617436 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.618947 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.632253 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.636928 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.641239 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.641400 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54d47\" (UniqueName: \"kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47\") pod \"nova-metadata-0\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.654801 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s7rg\" (UniqueName: \"kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg\") pod \"nova-api-0\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.663297 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.685942 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700333 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700379 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700417 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gzgh\" (UniqueName: \"kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700441 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700461 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bbs\" (UniqueName: \"kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700476 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700505 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbz2\" (UniqueName: \"kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700551 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700583 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700606 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700646 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.700674 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.706331 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.706533 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.707286 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.727822 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gzgh\" (UniqueName: \"kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh\") pod \"nova-scheduler-0\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.730060 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803221 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803506 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803557 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803608 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803630 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803658 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803677 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bbs\" (UniqueName: \"kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803699 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.803727 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbz2\" (UniqueName: \"kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.805099 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.805117 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.805613 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.806020 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.808768 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.810414 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.812426 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.816096 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.828581 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbz2\" (UniqueName: \"kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2\") pod \"nova-cell1-novncproxy-0\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.838921 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bbs\" (UniqueName: \"kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs\") pod \"dnsmasq-dns-865f5d856f-2ll65\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.870520 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:52 crc kubenswrapper[4751]: I1002 11:17:52.871212 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.207088 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-sq6r8"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.250437 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2nsmc"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.251894 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.254915 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.255286 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.262139 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2nsmc"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.319625 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.319711 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhwr\" (UniqueName: \"kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.319788 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.319822 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.334555 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.363604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sq6r8" event={"ID":"6092dc7b-23dd-4831-a5d7-ad61b6a5d237","Type":"ContainerStarted","Data":"809e98475834ab522efe090256a2fe9e96a78a05f82d9bd7744008067d7aef88"} Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.366049 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerStarted","Data":"df9629a7091a3e9e3f38cb08f9b783922497dfbd47549d823d003cec7bb2df5a"} Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.368601 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"df30039f-e3d2-43ce-aab5-050e2881c224","Type":"ContainerStarted","Data":"82aa9280ea289617ee4e316cca1cddf483f607b6e87705f81543cf1aa69af475"} Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.369132 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.393837 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9584610759999999 podStartE2EDuration="2.393812524s" podCreationTimestamp="2025-10-02 11:17:51 +0000 UTC" firstStartedPulling="2025-10-02 11:17:52.242193489 +0000 UTC m=+1554.296419939" lastFinishedPulling="2025-10-02 11:17:52.677544937 +0000 UTC m=+1554.731771387" observedRunningTime="2025-10-02 11:17:53.390399962 +0000 UTC m=+1555.444626412" watchObservedRunningTime="2025-10-02 11:17:53.393812524 +0000 UTC m=+1555.448038994" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.421215 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.421309 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhwr\" (UniqueName: \"kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.421392 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.421424 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.426118 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.427787 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.430486 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.441862 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.444966 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhwr\" (UniqueName: \"kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr\") pod \"nova-cell1-conductor-db-sync-2nsmc\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.449343 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.500836 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.543100 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.575500 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.588808 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:17:53 crc kubenswrapper[4751]: W1002 11:17:53.598206 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc58d30ab_0099_48d9_a2be_15c630e1179d.slice/crio-00f3ee00c24c7ed24baee1ca9528b9c6d2bac981f6f2c767f56d76b6e1b536ca WatchSource:0}: Error finding container 00f3ee00c24c7ed24baee1ca9528b9c6d2bac981f6f2c767f56d76b6e1b536ca: Status 404 returned error can't find the container with id 00f3ee00c24c7ed24baee1ca9528b9c6d2bac981f6f2c767f56d76b6e1b536ca Oct 02 11:17:53 crc kubenswrapper[4751]: I1002 11:17:53.604250 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:17:53 crc kubenswrapper[4751]: W1002 11:17:53.636665 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42567d4c_6c29_4c0c_b5d8_3c9c63f08326.slice/crio-2d44cc2cb65ae424ff0d5d714d665073bd5bcdb431d5e2f30c225278e691021d WatchSource:0}: Error finding container 2d44cc2cb65ae424ff0d5d714d665073bd5bcdb431d5e2f30c225278e691021d: Status 404 returned error can't find the container with id 2d44cc2cb65ae424ff0d5d714d665073bd5bcdb431d5e2f30c225278e691021d Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.082918 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2nsmc"] Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.386833 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sq6r8" event={"ID":"6092dc7b-23dd-4831-a5d7-ad61b6a5d237","Type":"ContainerStarted","Data":"20c9bca24ec08748601b40c0fa0461c7ba783625d330862673abe666392994b9"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.391707 4751 generic.go:334] "Generic (PLEG): container finished" podID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerID="4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968" exitCode=0 Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.391935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" event={"ID":"42567d4c-6c29-4c0c-b5d8-3c9c63f08326","Type":"ContainerDied","Data":"4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.391980 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" event={"ID":"42567d4c-6c29-4c0c-b5d8-3c9c63f08326","Type":"ContainerStarted","Data":"2d44cc2cb65ae424ff0d5d714d665073bd5bcdb431d5e2f30c225278e691021d"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.394876 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c58d30ab-0099-48d9-a2be-15c630e1179d","Type":"ContainerStarted","Data":"00f3ee00c24c7ed24baee1ca9528b9c6d2bac981f6f2c767f56d76b6e1b536ca"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.404620 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-sq6r8" podStartSLOduration=2.404599384 podStartE2EDuration="2.404599384s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:54.403792382 +0000 UTC m=+1556.458018832" watchObservedRunningTime="2025-10-02 11:17:54.404599384 +0000 UTC m=+1556.458825834" Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.408251 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" event={"ID":"e375b8c4-4852-45d4-a6d3-d47afbe12c90","Type":"ContainerStarted","Data":"dc96115b5cd493c6a2f1188e0cd795d4e1dda9a3e2acd5092634f0880bf29bbc"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.408306 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" event={"ID":"e375b8c4-4852-45d4-a6d3-d47afbe12c90","Type":"ContainerStarted","Data":"86424156248da3def79a852f3cef75db797fb4b9406d215b12d51674e1c68e6c"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.419585 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4189c7b4-f54f-4162-b97c-d210219b2784","Type":"ContainerStarted","Data":"1e71085b1455931ec2231f575bc5dbc866c5c2d2ad2b9b51845ce10d9717ac05"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.432508 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerStarted","Data":"54694475e9498211a0b06f06e03d1096dbcb292c2d941539f07aad86e2ca4dbb"} Oct 02 11:17:54 crc kubenswrapper[4751]: I1002 11:17:54.443422 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" podStartSLOduration=1.443398851 podStartE2EDuration="1.443398851s" podCreationTimestamp="2025-10-02 11:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:54.438290635 +0000 UTC m=+1556.492517115" watchObservedRunningTime="2025-10-02 11:17:54.443398851 +0000 UTC m=+1556.497625301" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.389225 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.391564 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.399749 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.455669 4751 generic.go:334] "Generic (PLEG): container finished" podID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerID="54c487812acd03ecde14f6528ebf21dce4be46c91cc5b57bd8938a831d09711c" exitCode=0 Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.455952 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5qfmq" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="registry-server" containerID="cri-o://afd4b0956824898b295f5afee0cf8fde826c093f1d97c943c33b32874b110233" gracePeriod=2 Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.456368 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerDied","Data":"54c487812acd03ecde14f6528ebf21dce4be46c91cc5b57bd8938a831d09711c"} Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.480864 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b57hx\" (UniqueName: \"kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.480936 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.481000 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.582617 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.582995 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b57hx\" (UniqueName: \"kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.583047 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.584976 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.585024 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.606869 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b57hx\" (UniqueName: \"kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx\") pod \"redhat-marketplace-nksc9\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.774811 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:17:55 crc kubenswrapper[4751]: I1002 11:17:55.994525 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:56 crc kubenswrapper[4751]: I1002 11:17:56.024457 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:17:56 crc kubenswrapper[4751]: I1002 11:17:56.473586 4751 generic.go:334] "Generic (PLEG): container finished" podID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerID="afd4b0956824898b295f5afee0cf8fde826c093f1d97c943c33b32874b110233" exitCode=0 Oct 02 11:17:56 crc kubenswrapper[4751]: I1002 11:17:56.473905 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerDied","Data":"afd4b0956824898b295f5afee0cf8fde826c093f1d97c943c33b32874b110233"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.081230 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.124872 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx4w6\" (UniqueName: \"kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6\") pod \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.124976 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content\") pod \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.125022 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities\") pod \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\" (UID: \"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.127367 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities" (OuterVolumeSpecName: "utilities") pod "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" (UID: "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.133641 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6" (OuterVolumeSpecName: "kube-api-access-jx4w6") pod "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" (UID: "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0"). InnerVolumeSpecName "kube-api-access-jx4w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.182639 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" (UID: "5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.206475 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.227018 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx4w6\" (UniqueName: \"kubernetes.io/projected/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-kube-api-access-jx4w6\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.227055 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.227068 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.305008 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.327990 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328159 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjxt8\" (UniqueName: \"kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328223 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328261 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328335 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328360 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.328407 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data\") pod \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\" (UID: \"3f6e3870-b235-40c3-b99a-ff6cb81e9138\") " Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.332874 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.333734 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.340605 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts" (OuterVolumeSpecName: "scripts") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.341264 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8" (OuterVolumeSpecName: "kube-api-access-kjxt8") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "kube-api-access-kjxt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.430409 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.430448 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjxt8\" (UniqueName: \"kubernetes.io/projected/3f6e3870-b235-40c3-b99a-ff6cb81e9138-kube-api-access-kjxt8\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.430466 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.430479 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f6e3870-b235-40c3-b99a-ff6cb81e9138-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.449443 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.492241 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" event={"ID":"42567d4c-6c29-4c0c-b5d8-3c9c63f08326","Type":"ContainerStarted","Data":"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.493458 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.498244 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c58d30ab-0099-48d9-a2be-15c630e1179d","Type":"ContainerStarted","Data":"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.498413 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c58d30ab-0099-48d9-a2be-15c630e1179d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3" gracePeriod=30 Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.502940 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerStarted","Data":"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.520943 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" podStartSLOduration=5.520921099 podStartE2EDuration="5.520921099s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:17:57.509917244 +0000 UTC m=+1559.564143694" watchObservedRunningTime="2025-10-02 11:17:57.520921099 +0000 UTC m=+1559.575147569" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.523213 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5qfmq" event={"ID":"5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0","Type":"ContainerDied","Data":"eb5b83f5bfe8bc78b40686058bd86ea0ee314415dcbef306e6813c1c9409eef3"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.523395 4751 scope.go:117] "RemoveContainer" containerID="afd4b0956824898b295f5afee0cf8fde826c093f1d97c943c33b32874b110233" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.527554 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5qfmq" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.532832 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.533550 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4189c7b4-f54f-4162-b97c-d210219b2784","Type":"ContainerStarted","Data":"dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.544284 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerStarted","Data":"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.546461 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.31317924 podStartE2EDuration="5.546436911s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="2025-10-02 11:17:53.604822364 +0000 UTC m=+1555.659048814" lastFinishedPulling="2025-10-02 11:17:56.838080035 +0000 UTC m=+1558.892306485" observedRunningTime="2025-10-02 11:17:57.524032762 +0000 UTC m=+1559.578259212" watchObservedRunningTime="2025-10-02 11:17:57.546436911 +0000 UTC m=+1559.600663381" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.550960 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.567751 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.245393939 podStartE2EDuration="5.56773207s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="2025-10-02 11:17:53.515662811 +0000 UTC m=+1555.569889261" lastFinishedPulling="2025-10-02 11:17:56.838000942 +0000 UTC m=+1558.892227392" observedRunningTime="2025-10-02 11:17:57.558709249 +0000 UTC m=+1559.612935699" watchObservedRunningTime="2025-10-02 11:17:57.56773207 +0000 UTC m=+1559.621958520" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.570521 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data" (OuterVolumeSpecName: "config-data") pod "3f6e3870-b235-40c3-b99a-ff6cb81e9138" (UID: "3f6e3870-b235-40c3-b99a-ff6cb81e9138"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.571442 4751 scope.go:117] "RemoveContainer" containerID="7e81669d0112becb5f33491d84d37db7fae724ac04c662ba5d497a4746fdcce5" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.584408 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerStarted","Data":"becf67fac96eb3fa5f8982e9d044b99e05334b9b6d0da221a52e22848ff0eb58"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.592785 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f6e3870-b235-40c3-b99a-ff6cb81e9138","Type":"ContainerDied","Data":"354e77288956bdfc70f959f02349733a9d23a4afc60601af761e52483a79c55a"} Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.592842 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.620829 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.648772 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.648815 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f6e3870-b235-40c3-b99a-ff6cb81e9138-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.655474 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5qfmq"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.657064 4751 scope.go:117] "RemoveContainer" containerID="9e40b94370dad0395be96ec3a364aee3f70e9f8bd89234b4cd25df98ca5417df" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.680311 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.689354 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.706848 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707366 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="extract-content" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707383 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="extract-content" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707400 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-notification-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707406 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-notification-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707420 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="registry-server" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707426 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="registry-server" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707439 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-central-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707445 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-central-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707467 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="sg-core" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707472 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="sg-core" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707480 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="extract-utilities" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707510 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="extract-utilities" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.707521 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="proxy-httpd" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707526 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="proxy-httpd" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707753 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="proxy-httpd" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707772 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-notification-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707781 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="sg-core" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707808 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" containerName="registry-server" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.707827 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" containerName="ceilometer-central-agent" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.710096 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.714230 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.714598 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.714702 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.714736 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.737455 4751 scope.go:117] "RemoveContainer" containerID="770b8ba699fda9d0f0ce88975e8848a4271b6081881becca62ecd48af60a5c47" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751132 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751378 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751498 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751571 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751594 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9s6c\" (UniqueName: \"kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751647 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751699 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.751735 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.759756 4751 scope.go:117] "RemoveContainer" containerID="e3943a58580606ecbf7a04c6a558b6adf8c6251d88865d401f322accd4a58722" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.816635 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.849569 4751 scope.go:117] "RemoveContainer" containerID="54c487812acd03ecde14f6528ebf21dce4be46c91cc5b57bd8938a831d09711c" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853036 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853113 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853133 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9s6c\" (UniqueName: \"kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853321 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853376 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853407 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853460 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.853515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.854597 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.859723 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.863590 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.869694 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.871907 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.873323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.874361 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: E1002 11:17:57.876390 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e0ef4bb_bff2_4d47_bc75_0af7a4463fc0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f6e3870_b235_40c3_b99a_ff6cb81e9138.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-conmon-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f6e3870_b235_40c3_b99a_ff6cb81e9138.slice/crio-354e77288956bdfc70f959f02349733a9d23a4afc60601af761e52483a79c55a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8235e6_b608_498e_b353_4197657298ea.slice/crio-0dc5c2e4d04370436c04e58c21f856864c910ca761ed66418b00d26676c4bb9c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.879872 4751 scope.go:117] "RemoveContainer" containerID="efa444739f95ded125966a8e41ecf8abbe44868f49908672a62d78b1fdeea688" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.881648 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9s6c\" (UniqueName: \"kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:57 crc kubenswrapper[4751]: I1002 11:17:57.895023 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " pod="openstack/ceilometer-0" Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.044676 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.489938 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.605981 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8a42e63-2379-4384-9bea-f756390a1935" containerID="7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115" exitCode=0 Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.606073 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerDied","Data":"7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115"} Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.608080 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerStarted","Data":"96003922d481185750396f4310672af285bc506acdb843cad29320f894c63b96"} Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.614211 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerStarted","Data":"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36"} Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.619404 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerStarted","Data":"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c"} Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.619568 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-log" containerID="cri-o://15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" gracePeriod=30 Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.619726 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-metadata" containerID="cri-o://c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" gracePeriod=30 Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.654787 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.273874131 podStartE2EDuration="6.654761968s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="2025-10-02 11:17:53.455883273 +0000 UTC m=+1555.510109723" lastFinishedPulling="2025-10-02 11:17:56.83677111 +0000 UTC m=+1558.890997560" observedRunningTime="2025-10-02 11:17:58.647193326 +0000 UTC m=+1560.701419776" watchObservedRunningTime="2025-10-02 11:17:58.654761968 +0000 UTC m=+1560.708988428" Oct 02 11:17:58 crc kubenswrapper[4751]: I1002 11:17:58.672848 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.173157809 podStartE2EDuration="6.672829091s" podCreationTimestamp="2025-10-02 11:17:52 +0000 UTC" firstStartedPulling="2025-10-02 11:17:53.33720156 +0000 UTC m=+1555.391428010" lastFinishedPulling="2025-10-02 11:17:56.836872842 +0000 UTC m=+1558.891099292" observedRunningTime="2025-10-02 11:17:58.663666746 +0000 UTC m=+1560.717893196" watchObservedRunningTime="2025-10-02 11:17:58.672829091 +0000 UTC m=+1560.727055541" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.138630 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.182149 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data\") pod \"d638d372-33da-4a91-af83-fb090f7bb54d\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.182246 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs\") pod \"d638d372-33da-4a91-af83-fb090f7bb54d\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.182664 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle\") pod \"d638d372-33da-4a91-af83-fb090f7bb54d\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.182593 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs" (OuterVolumeSpecName: "logs") pod "d638d372-33da-4a91-af83-fb090f7bb54d" (UID: "d638d372-33da-4a91-af83-fb090f7bb54d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.183094 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54d47\" (UniqueName: \"kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47\") pod \"d638d372-33da-4a91-af83-fb090f7bb54d\" (UID: \"d638d372-33da-4a91-af83-fb090f7bb54d\") " Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.183868 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d638d372-33da-4a91-af83-fb090f7bb54d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.191744 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47" (OuterVolumeSpecName: "kube-api-access-54d47") pod "d638d372-33da-4a91-af83-fb090f7bb54d" (UID: "d638d372-33da-4a91-af83-fb090f7bb54d"). InnerVolumeSpecName "kube-api-access-54d47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.228655 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d638d372-33da-4a91-af83-fb090f7bb54d" (UID: "d638d372-33da-4a91-af83-fb090f7bb54d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.231670 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data" (OuterVolumeSpecName: "config-data") pod "d638d372-33da-4a91-af83-fb090f7bb54d" (UID: "d638d372-33da-4a91-af83-fb090f7bb54d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.287384 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54d47\" (UniqueName: \"kubernetes.io/projected/d638d372-33da-4a91-af83-fb090f7bb54d-kube-api-access-54d47\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.287419 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.287429 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d638d372-33da-4a91-af83-fb090f7bb54d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:17:59 crc kubenswrapper[4751]: E1002 11:17:59.534636 4751 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b56dd1112ce7097fb00d35f3a85cd6924b85516e402f3f29b9e89a4dabe8e178/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b56dd1112ce7097fb00d35f3a85cd6924b85516e402f3f29b9e89a4dabe8e178/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_community-operators-w7lj5_cb8235e6-b608-498e-b353-4197657298ea/registry-server/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_community-operators-w7lj5_cb8235e6-b608-498e-b353-4197657298ea/registry-server/0.log: no such file or directory Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.566120 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f6e3870-b235-40c3-b99a-ff6cb81e9138" path="/var/lib/kubelet/pods/3f6e3870-b235-40c3-b99a-ff6cb81e9138/volumes" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.566998 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0" path="/var/lib/kubelet/pods/5e0ef4bb-bff2-4d47-bc75-0af7a4463fc0/volumes" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.629063 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerStarted","Data":"ce301b246e99c191f9083946459cef1b1fbe155015e8dbafca08ea162a31e232"} Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632832 4751 generic.go:334] "Generic (PLEG): container finished" podID="d638d372-33da-4a91-af83-fb090f7bb54d" containerID="c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" exitCode=0 Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632856 4751 generic.go:334] "Generic (PLEG): container finished" podID="d638d372-33da-4a91-af83-fb090f7bb54d" containerID="15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" exitCode=143 Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632866 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerDied","Data":"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c"} Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632898 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerDied","Data":"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57"} Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632910 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d638d372-33da-4a91-af83-fb090f7bb54d","Type":"ContainerDied","Data":"54694475e9498211a0b06f06e03d1096dbcb292c2d941539f07aad86e2ca4dbb"} Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632930 4751 scope.go:117] "RemoveContainer" containerID="c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.632942 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.635875 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8a42e63-2379-4384-9bea-f756390a1935" containerID="548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7" exitCode=0 Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.637006 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerDied","Data":"548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7"} Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.673766 4751 scope.go:117] "RemoveContainer" containerID="15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.686144 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.696933 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.707503 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:59 crc kubenswrapper[4751]: E1002 11:17:59.707884 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-log" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.707901 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-log" Oct 02 11:17:59 crc kubenswrapper[4751]: E1002 11:17:59.707942 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-metadata" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.707951 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-metadata" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.708139 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-metadata" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.708180 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" containerName="nova-metadata-log" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.709078 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.712478 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.712743 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.745436 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.749110 4751 scope.go:117] "RemoveContainer" containerID="c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" Oct 02 11:17:59 crc kubenswrapper[4751]: E1002 11:17:59.749731 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c\": container with ID starting with c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c not found: ID does not exist" containerID="c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.749761 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c"} err="failed to get container status \"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c\": rpc error: code = NotFound desc = could not find container \"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c\": container with ID starting with c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c not found: ID does not exist" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.749781 4751 scope.go:117] "RemoveContainer" containerID="15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" Oct 02 11:17:59 crc kubenswrapper[4751]: E1002 11:17:59.756412 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57\": container with ID starting with 15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57 not found: ID does not exist" containerID="15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.756461 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57"} err="failed to get container status \"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57\": rpc error: code = NotFound desc = could not find container \"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57\": container with ID starting with 15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57 not found: ID does not exist" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.756529 4751 scope.go:117] "RemoveContainer" containerID="c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.758290 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c"} err="failed to get container status \"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c\": rpc error: code = NotFound desc = could not find container \"c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c\": container with ID starting with c81cdf1068ef1601cfb3794a377a1f1380a5e70536fe56b3d2081b9bf7c8236c not found: ID does not exist" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.758310 4751 scope.go:117] "RemoveContainer" containerID="15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.758857 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57"} err="failed to get container status \"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57\": rpc error: code = NotFound desc = could not find container \"15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57\": container with ID starting with 15ada889fe41093d8c8f64330dec06c1f9c924ef3e5ddf1bd78366b2b6b0ce57 not found: ID does not exist" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.797361 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddhc6\" (UniqueName: \"kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.797416 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.797445 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.797603 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.797637 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.899255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.899306 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.899372 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.899396 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.899484 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddhc6\" (UniqueName: \"kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.900671 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.902933 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.907875 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.913356 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:17:59 crc kubenswrapper[4751]: I1002 11:17:59.916312 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddhc6\" (UniqueName: \"kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6\") pod \"nova-metadata-0\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " pod="openstack/nova-metadata-0" Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.029704 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.501606 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.661529 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerStarted","Data":"124bddfa11944597d4f7a60eb4433125f1170bea975e172a384ba3785ac596ec"} Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.671054 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerStarted","Data":"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f"} Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.681688 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerStarted","Data":"e4995510d541c879a0992ef0e05126af02aead27d8708e9e0daba361c76661b4"} Oct 02 11:18:00 crc kubenswrapper[4751]: I1002 11:18:00.703913 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nksc9" podStartSLOduration=4.0027049 podStartE2EDuration="5.703887655s" podCreationTimestamp="2025-10-02 11:17:55 +0000 UTC" firstStartedPulling="2025-10-02 11:17:58.60766959 +0000 UTC m=+1560.661896040" lastFinishedPulling="2025-10-02 11:18:00.308852345 +0000 UTC m=+1562.363078795" observedRunningTime="2025-10-02 11:18:00.69473152 +0000 UTC m=+1562.748957980" watchObservedRunningTime="2025-10-02 11:18:00.703887655 +0000 UTC m=+1562.758114105" Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.506947 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.507310 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.563134 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d638d372-33da-4a91-af83-fb090f7bb54d" path="/var/lib/kubelet/pods/d638d372-33da-4a91-af83-fb090f7bb54d/volumes" Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.713548 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerStarted","Data":"9a086cc41a2751992b4d40879db5a227aecde8b5e959356e7f30d81041832ceb"} Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.725734 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerStarted","Data":"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1"} Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.726013 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerStarted","Data":"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad"} Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.746282 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 11:18:01 crc kubenswrapper[4751]: I1002 11:18:01.760207 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.760156201 podStartE2EDuration="2.760156201s" podCreationTimestamp="2025-10-02 11:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:01.753554355 +0000 UTC m=+1563.807780815" watchObservedRunningTime="2025-10-02 11:18:01.760156201 +0000 UTC m=+1563.814382651" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.686922 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.687241 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.741201 4751 generic.go:334] "Generic (PLEG): container finished" podID="6092dc7b-23dd-4831-a5d7-ad61b6a5d237" containerID="20c9bca24ec08748601b40c0fa0461c7ba783625d330862673abe666392994b9" exitCode=0 Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.741303 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sq6r8" event={"ID":"6092dc7b-23dd-4831-a5d7-ad61b6a5d237","Type":"ContainerDied","Data":"20c9bca24ec08748601b40c0fa0461c7ba783625d330862673abe666392994b9"} Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.817545 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.868896 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.876354 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.975372 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:18:02 crc kubenswrapper[4751]: I1002 11:18:02.975609 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="dnsmasq-dns" containerID="cri-o://ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8" gracePeriod=10 Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.604619 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681030 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jht8l\" (UniqueName: \"kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681139 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681229 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681258 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681286 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.681468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb\") pod \"189cc954-c9db-462c-a75d-b7825503e11a\" (UID: \"189cc954-c9db-462c-a75d-b7825503e11a\") " Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.696495 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l" (OuterVolumeSpecName: "kube-api-access-jht8l") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "kube-api-access-jht8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.734749 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.757847 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config" (OuterVolumeSpecName: "config") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.764268 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerStarted","Data":"c67cd3ef1c4fe61ca5288a671379d49fa8971ba5c1aa63aebf51fa029ec17f8a"} Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.765481 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.775645 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.776420 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.776578 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" event={"ID":"189cc954-c9db-462c-a75d-b7825503e11a","Type":"ContainerDied","Data":"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8"} Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.776763 4751 scope.go:117] "RemoveContainer" containerID="ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.776350 4751 generic.go:334] "Generic (PLEG): container finished" podID="189cc954-c9db-462c-a75d-b7825503e11a" containerID="ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8" exitCode=0 Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.777926 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-lfjcn" event={"ID":"189cc954-c9db-462c-a75d-b7825503e11a","Type":"ContainerDied","Data":"d192b176f499d8793cd758808525231af9143c1cff314c84795e4e760053c1db"} Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.784760 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jht8l\" (UniqueName: \"kubernetes.io/projected/189cc954-c9db-462c-a75d-b7825503e11a-kube-api-access-jht8l\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.784919 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.799816 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6958371789999998 podStartE2EDuration="6.799797665s" podCreationTimestamp="2025-10-02 11:17:57 +0000 UTC" firstStartedPulling="2025-10-02 11:17:58.503098224 +0000 UTC m=+1560.557324684" lastFinishedPulling="2025-10-02 11:18:02.60705872 +0000 UTC m=+1564.661285170" observedRunningTime="2025-10-02 11:18:03.797710919 +0000 UTC m=+1565.851937389" watchObservedRunningTime="2025-10-02 11:18:03.799797665 +0000 UTC m=+1565.854024115" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.808875 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.816066 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.819819 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.823372 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "189cc954-c9db-462c-a75d-b7825503e11a" (UID: "189cc954-c9db-462c-a75d-b7825503e11a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.834337 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.838467 4751 scope.go:117] "RemoveContainer" containerID="018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.872798 4751 scope.go:117] "RemoveContainer" containerID="ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8" Oct 02 11:18:03 crc kubenswrapper[4751]: E1002 11:18:03.873117 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8\": container with ID starting with ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8 not found: ID does not exist" containerID="ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.873153 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8"} err="failed to get container status \"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8\": rpc error: code = NotFound desc = could not find container \"ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8\": container with ID starting with ef1eb720a0223752419b3e2ea71ea0e1942ad8a16ab78ac9b2e4fba8a77b38c8 not found: ID does not exist" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.873188 4751 scope.go:117] "RemoveContainer" containerID="018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513" Oct 02 11:18:03 crc kubenswrapper[4751]: E1002 11:18:03.875808 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513\": container with ID starting with 018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513 not found: ID does not exist" containerID="018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.875848 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513"} err="failed to get container status \"018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513\": rpc error: code = NotFound desc = could not find container \"018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513\": container with ID starting with 018a65ce9b602d8655dd470006c2011b6f1a3200ca48ae6d4102c8c44f710513 not found: ID does not exist" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.889122 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.889151 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.889161 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:03 crc kubenswrapper[4751]: I1002 11:18:03.889186 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/189cc954-c9db-462c-a75d-b7825503e11a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.218015 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.283759 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.295556 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xgrf\" (UniqueName: \"kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf\") pod \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.295869 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data\") pod \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.295918 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle\") pod \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.295985 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts\") pod \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\" (UID: \"6092dc7b-23dd-4831-a5d7-ad61b6a5d237\") " Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.305421 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-lfjcn"] Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.305438 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts" (OuterVolumeSpecName: "scripts") pod "6092dc7b-23dd-4831-a5d7-ad61b6a5d237" (UID: "6092dc7b-23dd-4831-a5d7-ad61b6a5d237"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.319023 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf" (OuterVolumeSpecName: "kube-api-access-5xgrf") pod "6092dc7b-23dd-4831-a5d7-ad61b6a5d237" (UID: "6092dc7b-23dd-4831-a5d7-ad61b6a5d237"). InnerVolumeSpecName "kube-api-access-5xgrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.379369 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data" (OuterVolumeSpecName: "config-data") pod "6092dc7b-23dd-4831-a5d7-ad61b6a5d237" (UID: "6092dc7b-23dd-4831-a5d7-ad61b6a5d237"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.397850 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.397880 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xgrf\" (UniqueName: \"kubernetes.io/projected/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-kube-api-access-5xgrf\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.397891 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.457415 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6092dc7b-23dd-4831-a5d7-ad61b6a5d237" (UID: "6092dc7b-23dd-4831-a5d7-ad61b6a5d237"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.500334 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6092dc7b-23dd-4831-a5d7-ad61b6a5d237-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.789907 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sq6r8" event={"ID":"6092dc7b-23dd-4831-a5d7-ad61b6a5d237","Type":"ContainerDied","Data":"809e98475834ab522efe090256a2fe9e96a78a05f82d9bd7744008067d7aef88"} Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.789966 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sq6r8" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.789960 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="809e98475834ab522efe090256a2fe9e96a78a05f82d9bd7744008067d7aef88" Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.791488 4751 generic.go:334] "Generic (PLEG): container finished" podID="e375b8c4-4852-45d4-a6d3-d47afbe12c90" containerID="dc96115b5cd493c6a2f1188e0cd795d4e1dda9a3e2acd5092634f0880bf29bbc" exitCode=0 Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.791532 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" event={"ID":"e375b8c4-4852-45d4-a6d3-d47afbe12c90","Type":"ContainerDied","Data":"dc96115b5cd493c6a2f1188e0cd795d4e1dda9a3e2acd5092634f0880bf29bbc"} Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.941882 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.942104 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-log" containerID="cri-o://1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877" gracePeriod=30 Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.942508 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-api" containerID="cri-o://3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36" gracePeriod=30 Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.951730 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.990891 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.991226 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-log" containerID="cri-o://758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" gracePeriod=30 Oct 02 11:18:04 crc kubenswrapper[4751]: I1002 11:18:04.991297 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-metadata" containerID="cri-o://fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" gracePeriod=30 Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.030937 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.034289 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.575612 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189cc954-c9db-462c-a75d-b7825503e11a" path="/var/lib/kubelet/pods/189cc954-c9db-462c-a75d-b7825503e11a/volumes" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.689630 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.725576 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs\") pod \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.725654 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs\") pod \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.725708 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data\") pod \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.725758 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle\") pod \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.725842 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddhc6\" (UniqueName: \"kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6\") pod \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\" (UID: \"0591cb81-3c5d-4c93-8a9c-88000aa3df5c\") " Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.727592 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs" (OuterVolumeSpecName: "logs") pod "0591cb81-3c5d-4c93-8a9c-88000aa3df5c" (UID: "0591cb81-3c5d-4c93-8a9c-88000aa3df5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.735956 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6" (OuterVolumeSpecName: "kube-api-access-ddhc6") pod "0591cb81-3c5d-4c93-8a9c-88000aa3df5c" (UID: "0591cb81-3c5d-4c93-8a9c-88000aa3df5c"). InnerVolumeSpecName "kube-api-access-ddhc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.752484 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0591cb81-3c5d-4c93-8a9c-88000aa3df5c" (UID: "0591cb81-3c5d-4c93-8a9c-88000aa3df5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.761726 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data" (OuterVolumeSpecName: "config-data") pod "0591cb81-3c5d-4c93-8a9c-88000aa3df5c" (UID: "0591cb81-3c5d-4c93-8a9c-88000aa3df5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.775441 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.775718 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.806292 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0591cb81-3c5d-4c93-8a9c-88000aa3df5c" (UID: "0591cb81-3c5d-4c93-8a9c-88000aa3df5c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.814394 4751 generic.go:334] "Generic (PLEG): container finished" podID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerID="fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" exitCode=0 Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.814651 4751 generic.go:334] "Generic (PLEG): container finished" podID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerID="758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" exitCode=143 Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.814627 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.814532 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerDied","Data":"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1"} Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.825628 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerDied","Data":"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad"} Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.825790 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0591cb81-3c5d-4c93-8a9c-88000aa3df5c","Type":"ContainerDied","Data":"124bddfa11944597d4f7a60eb4433125f1170bea975e172a384ba3785ac596ec"} Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.825921 4751 scope.go:117] "RemoveContainer" containerID="fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.828379 4751 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.828398 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.828423 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.828435 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.828445 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddhc6\" (UniqueName: \"kubernetes.io/projected/0591cb81-3c5d-4c93-8a9c-88000aa3df5c-kube-api-access-ddhc6\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.840370 4751 generic.go:334] "Generic (PLEG): container finished" podID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerID="1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877" exitCode=143 Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.840568 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" containerName="nova-scheduler-scheduler" containerID="cri-o://dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" gracePeriod=30 Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.840757 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerDied","Data":"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877"} Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.840878 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.898457 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.900769 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.907139 4751 scope.go:117] "RemoveContainer" containerID="758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.914245 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.943990 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.944500 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-metadata" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944522 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-metadata" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.944554 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-log" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944563 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-log" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.944583 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="init" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944592 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="init" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.944625 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6092dc7b-23dd-4831-a5d7-ad61b6a5d237" containerName="nova-manage" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944634 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6092dc7b-23dd-4831-a5d7-ad61b6a5d237" containerName="nova-manage" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.944645 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="dnsmasq-dns" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944653 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="dnsmasq-dns" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944861 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-metadata" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944879 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6092dc7b-23dd-4831-a5d7-ad61b6a5d237" containerName="nova-manage" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944890 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="189cc954-c9db-462c-a75d-b7825503e11a" containerName="dnsmasq-dns" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.944922 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" containerName="nova-metadata-log" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.946254 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.949585 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.951087 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.985908 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.986209 4751 scope.go:117] "RemoveContainer" containerID="fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.989309 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1\": container with ID starting with fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1 not found: ID does not exist" containerID="fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.989359 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1"} err="failed to get container status \"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1\": rpc error: code = NotFound desc = could not find container \"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1\": container with ID starting with fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1 not found: ID does not exist" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.989395 4751 scope.go:117] "RemoveContainer" containerID="758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" Oct 02 11:18:05 crc kubenswrapper[4751]: E1002 11:18:05.992533 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad\": container with ID starting with 758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad not found: ID does not exist" containerID="758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.992574 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad"} err="failed to get container status \"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad\": rpc error: code = NotFound desc = could not find container \"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad\": container with ID starting with 758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad not found: ID does not exist" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.992597 4751 scope.go:117] "RemoveContainer" containerID="fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.993736 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1"} err="failed to get container status \"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1\": rpc error: code = NotFound desc = could not find container \"fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1\": container with ID starting with fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1 not found: ID does not exist" Oct 02 11:18:05 crc kubenswrapper[4751]: I1002 11:18:05.993783 4751 scope.go:117] "RemoveContainer" containerID="758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:05.997526 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad"} err="failed to get container status \"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad\": rpc error: code = NotFound desc = could not find container \"758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad\": container with ID starting with 758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad not found: ID does not exist" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.035503 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.035562 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4ssr\" (UniqueName: \"kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.035601 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.035697 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.035782 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.086738 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.137389 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.137467 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.137544 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.137565 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4ssr\" (UniqueName: \"kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.137589 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.138282 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.141488 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.141571 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.142140 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.160233 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4ssr\" (UniqueName: \"kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr\") pod \"nova-metadata-0\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.267394 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.378382 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.544024 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts\") pod \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.544160 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dhwr\" (UniqueName: \"kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr\") pod \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.544300 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle\") pod \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.544424 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data\") pod \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\" (UID: \"e375b8c4-4852-45d4-a6d3-d47afbe12c90\") " Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.553227 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts" (OuterVolumeSpecName: "scripts") pod "e375b8c4-4852-45d4-a6d3-d47afbe12c90" (UID: "e375b8c4-4852-45d4-a6d3-d47afbe12c90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.553514 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr" (OuterVolumeSpecName: "kube-api-access-7dhwr") pod "e375b8c4-4852-45d4-a6d3-d47afbe12c90" (UID: "e375b8c4-4852-45d4-a6d3-d47afbe12c90"). InnerVolumeSpecName "kube-api-access-7dhwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.586062 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e375b8c4-4852-45d4-a6d3-d47afbe12c90" (UID: "e375b8c4-4852-45d4-a6d3-d47afbe12c90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.587428 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data" (OuterVolumeSpecName: "config-data") pod "e375b8c4-4852-45d4-a6d3-d47afbe12c90" (UID: "e375b8c4-4852-45d4-a6d3-d47afbe12c90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.646886 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.646925 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dhwr\" (UniqueName: \"kubernetes.io/projected/e375b8c4-4852-45d4-a6d3-d47afbe12c90-kube-api-access-7dhwr\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.646940 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.646950 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e375b8c4-4852-45d4-a6d3-d47afbe12c90-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.765839 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.857846 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" event={"ID":"e375b8c4-4852-45d4-a6d3-d47afbe12c90","Type":"ContainerDied","Data":"86424156248da3def79a852f3cef75db797fb4b9406d215b12d51674e1c68e6c"} Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.859609 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86424156248da3def79a852f3cef75db797fb4b9406d215b12d51674e1c68e6c" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.859902 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2nsmc" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.864848 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerStarted","Data":"19e2dec392aa9e70eac6f5aa099170e99b5f55cac7bfb8e726652e9c54cabc80"} Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.911294 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 11:18:06 crc kubenswrapper[4751]: E1002 11:18:06.911779 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e375b8c4-4852-45d4-a6d3-d47afbe12c90" containerName="nova-cell1-conductor-db-sync" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.911803 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e375b8c4-4852-45d4-a6d3-d47afbe12c90" containerName="nova-cell1-conductor-db-sync" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.912107 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="e375b8c4-4852-45d4-a6d3-d47afbe12c90" containerName="nova-cell1-conductor-db-sync" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.912915 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.918640 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.929072 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.955331 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.955649 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:06 crc kubenswrapper[4751]: I1002 11:18:06.955773 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmvwc\" (UniqueName: \"kubernetes.io/projected/03553bb2-dd3a-4362-946d-4379d45afa2d-kube-api-access-kmvwc\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.056765 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.057446 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmvwc\" (UniqueName: \"kubernetes.io/projected/03553bb2-dd3a-4362-946d-4379d45afa2d-kube-api-access-kmvwc\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.058109 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.060712 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.061023 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03553bb2-dd3a-4362-946d-4379d45afa2d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.076430 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmvwc\" (UniqueName: \"kubernetes.io/projected/03553bb2-dd3a-4362-946d-4379d45afa2d-kube-api-access-kmvwc\") pod \"nova-cell1-conductor-0\" (UID: \"03553bb2-dd3a-4362-946d-4379d45afa2d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.236839 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.563130 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0591cb81-3c5d-4c93-8a9c-88000aa3df5c" path="/var/lib/kubelet/pods/0591cb81-3c5d-4c93-8a9c-88000aa3df5c/volumes" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.692810 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 11:18:07 crc kubenswrapper[4751]: E1002 11:18:07.819350 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 11:18:07 crc kubenswrapper[4751]: E1002 11:18:07.821536 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 11:18:07 crc kubenswrapper[4751]: E1002 11:18:07.826660 4751 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 11:18:07 crc kubenswrapper[4751]: E1002 11:18:07.826737 4751 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" containerName="nova-scheduler-scheduler" Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.876281 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerStarted","Data":"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4"} Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.876329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerStarted","Data":"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206"} Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.881626 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"03553bb2-dd3a-4362-946d-4379d45afa2d","Type":"ContainerStarted","Data":"18986496fa4c37f5478a1f2568fbdc2d438d88e130d1b4122d9546edb1d42f4f"} Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.881847 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nksc9" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="registry-server" containerID="cri-o://d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f" gracePeriod=2 Oct 02 11:18:07 crc kubenswrapper[4751]: I1002 11:18:07.914815 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.914791315 podStartE2EDuration="2.914791315s" podCreationTimestamp="2025-10-02 11:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:07.902890457 +0000 UTC m=+1569.957116907" watchObservedRunningTime="2025-10-02 11:18:07.914791315 +0000 UTC m=+1569.969017765" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.462502 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.591805 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities\") pod \"e8a42e63-2379-4384-9bea-f756390a1935\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.591875 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content\") pod \"e8a42e63-2379-4384-9bea-f756390a1935\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.591947 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b57hx\" (UniqueName: \"kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx\") pod \"e8a42e63-2379-4384-9bea-f756390a1935\" (UID: \"e8a42e63-2379-4384-9bea-f756390a1935\") " Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.592762 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities" (OuterVolumeSpecName: "utilities") pod "e8a42e63-2379-4384-9bea-f756390a1935" (UID: "e8a42e63-2379-4384-9bea-f756390a1935"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.599956 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx" (OuterVolumeSpecName: "kube-api-access-b57hx") pod "e8a42e63-2379-4384-9bea-f756390a1935" (UID: "e8a42e63-2379-4384-9bea-f756390a1935"). InnerVolumeSpecName "kube-api-access-b57hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.610080 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8a42e63-2379-4384-9bea-f756390a1935" (UID: "e8a42e63-2379-4384-9bea-f756390a1935"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.695443 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.695479 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8a42e63-2379-4384-9bea-f756390a1935-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.695492 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b57hx\" (UniqueName: \"kubernetes.io/projected/e8a42e63-2379-4384-9bea-f756390a1935-kube-api-access-b57hx\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.894094 4751 generic.go:334] "Generic (PLEG): container finished" podID="e8a42e63-2379-4384-9bea-f756390a1935" containerID="d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f" exitCode=0 Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.894174 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerDied","Data":"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f"} Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.894225 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nksc9" event={"ID":"e8a42e63-2379-4384-9bea-f756390a1935","Type":"ContainerDied","Data":"becf67fac96eb3fa5f8982e9d044b99e05334b9b6d0da221a52e22848ff0eb58"} Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.894244 4751 scope.go:117] "RemoveContainer" containerID="d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.894398 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nksc9" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.905329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"03553bb2-dd3a-4362-946d-4379d45afa2d","Type":"ContainerStarted","Data":"585e9eebd635f6e72cb8de0781463e77a50e87127dee4224b5b2d34bd0604a0c"} Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.930004 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.929978143 podStartE2EDuration="2.929978143s" podCreationTimestamp="2025-10-02 11:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:08.925340829 +0000 UTC m=+1570.979567299" watchObservedRunningTime="2025-10-02 11:18:08.929978143 +0000 UTC m=+1570.984204593" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.946655 4751 scope.go:117] "RemoveContainer" containerID="548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7" Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.951328 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.961007 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nksc9"] Oct 02 11:18:08 crc kubenswrapper[4751]: I1002 11:18:08.983687 4751 scope.go:117] "RemoveContainer" containerID="7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.029743 4751 scope.go:117] "RemoveContainer" containerID="d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f" Oct 02 11:18:09 crc kubenswrapper[4751]: E1002 11:18:09.030473 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f\": container with ID starting with d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f not found: ID does not exist" containerID="d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.030579 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f"} err="failed to get container status \"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f\": rpc error: code = NotFound desc = could not find container \"d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f\": container with ID starting with d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f not found: ID does not exist" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.030657 4751 scope.go:117] "RemoveContainer" containerID="548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7" Oct 02 11:18:09 crc kubenswrapper[4751]: E1002 11:18:09.030992 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7\": container with ID starting with 548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7 not found: ID does not exist" containerID="548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.031031 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7"} err="failed to get container status \"548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7\": rpc error: code = NotFound desc = could not find container \"548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7\": container with ID starting with 548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7 not found: ID does not exist" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.031051 4751 scope.go:117] "RemoveContainer" containerID="7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115" Oct 02 11:18:09 crc kubenswrapper[4751]: E1002 11:18:09.031601 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115\": container with ID starting with 7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115 not found: ID does not exist" containerID="7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.031630 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115"} err="failed to get container status \"7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115\": rpc error: code = NotFound desc = could not find container \"7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115\": container with ID starting with 7cc2aaa29a18e1fe696a2e434233a8f7f2e8e08e620a9ea566237fa92acdd115 not found: ID does not exist" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.567879 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8a42e63-2379-4384-9bea-f756390a1935" path="/var/lib/kubelet/pods/e8a42e63-2379-4384-9bea-f756390a1935/volumes" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.809223 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.945259 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s7rg\" (UniqueName: \"kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg\") pod \"5ca6623c-2f50-48ff-a82c-bc822c443f30\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.945347 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs\") pod \"5ca6623c-2f50-48ff-a82c-bc822c443f30\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.945430 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data\") pod \"5ca6623c-2f50-48ff-a82c-bc822c443f30\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.945501 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle\") pod \"5ca6623c-2f50-48ff-a82c-bc822c443f30\" (UID: \"5ca6623c-2f50-48ff-a82c-bc822c443f30\") " Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.953904 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg" (OuterVolumeSpecName: "kube-api-access-7s7rg") pod "5ca6623c-2f50-48ff-a82c-bc822c443f30" (UID: "5ca6623c-2f50-48ff-a82c-bc822c443f30"). InnerVolumeSpecName "kube-api-access-7s7rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.960496 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs" (OuterVolumeSpecName: "logs") pod "5ca6623c-2f50-48ff-a82c-bc822c443f30" (UID: "5ca6623c-2f50-48ff-a82c-bc822c443f30"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.982888 4751 generic.go:334] "Generic (PLEG): container finished" podID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerID="3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36" exitCode=0 Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.983007 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerDied","Data":"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36"} Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.983039 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ca6623c-2f50-48ff-a82c-bc822c443f30","Type":"ContainerDied","Data":"df9629a7091a3e9e3f38cb08f9b783922497dfbd47549d823d003cec7bb2df5a"} Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.983057 4751 scope.go:117] "RemoveContainer" containerID="3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.983245 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.991672 4751 generic.go:334] "Generic (PLEG): container finished" podID="4189c7b4-f54f-4162-b97c-d210219b2784" containerID="dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" exitCode=0 Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.992694 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data" (OuterVolumeSpecName: "config-data") pod "5ca6623c-2f50-48ff-a82c-bc822c443f30" (UID: "5ca6623c-2f50-48ff-a82c-bc822c443f30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.992788 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4189c7b4-f54f-4162-b97c-d210219b2784","Type":"ContainerDied","Data":"dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046"} Oct 02 11:18:09 crc kubenswrapper[4751]: I1002 11:18:09.992843 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.001263 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ca6623c-2f50-48ff-a82c-bc822c443f30" (UID: "5ca6623c-2f50-48ff-a82c-bc822c443f30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.017373 4751 scope.go:117] "RemoveContainer" containerID="1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.047117 4751 scope.go:117] "RemoveContainer" containerID="3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.047533 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s7rg\" (UniqueName: \"kubernetes.io/projected/5ca6623c-2f50-48ff-a82c-bc822c443f30-kube-api-access-7s7rg\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.047563 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ca6623c-2f50-48ff-a82c-bc822c443f30-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.047574 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.047584 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ca6623c-2f50-48ff-a82c-bc822c443f30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.049113 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36\": container with ID starting with 3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36 not found: ID does not exist" containerID="3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.049172 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36"} err="failed to get container status \"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36\": rpc error: code = NotFound desc = could not find container \"3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36\": container with ID starting with 3aa89f53a2d37b12f3b42ab75ad0b9d3adad7684117926b688040b70d8821d36 not found: ID does not exist" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.049201 4751 scope.go:117] "RemoveContainer" containerID="1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.051062 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877\": container with ID starting with 1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877 not found: ID does not exist" containerID="1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.051105 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877"} err="failed to get container status \"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877\": rpc error: code = NotFound desc = could not find container \"1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877\": container with ID starting with 1cffc873bb137f45dc4833eee7f0431dd5148dc222e817e99b6d077f4e4fe877 not found: ID does not exist" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.340040 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.359647 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.371826 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409272 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409729 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-api" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409756 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-api" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409779 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="extract-utilities" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409788 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="extract-utilities" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409818 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="extract-content" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409827 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="extract-content" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409841 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" containerName="nova-scheduler-scheduler" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409850 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" containerName="nova-scheduler-scheduler" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409869 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-log" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409877 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-log" Oct 02 11:18:10 crc kubenswrapper[4751]: E1002 11:18:10.409887 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="registry-server" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.409894 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="registry-server" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.410132 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-api" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.410157 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" containerName="nova-api-log" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.410190 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" containerName="nova-scheduler-scheduler" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.410205 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8a42e63-2379-4384-9bea-f756390a1935" containerName="registry-server" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.411425 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.413743 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.427425 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.452800 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle\") pod \"4189c7b4-f54f-4162-b97c-d210219b2784\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.452885 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data\") pod \"4189c7b4-f54f-4162-b97c-d210219b2784\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.452984 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gzgh\" (UniqueName: \"kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh\") pod \"4189c7b4-f54f-4162-b97c-d210219b2784\" (UID: \"4189c7b4-f54f-4162-b97c-d210219b2784\") " Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.456875 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh" (OuterVolumeSpecName: "kube-api-access-7gzgh") pod "4189c7b4-f54f-4162-b97c-d210219b2784" (UID: "4189c7b4-f54f-4162-b97c-d210219b2784"). InnerVolumeSpecName "kube-api-access-7gzgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.495484 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4189c7b4-f54f-4162-b97c-d210219b2784" (UID: "4189c7b4-f54f-4162-b97c-d210219b2784"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.523649 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data" (OuterVolumeSpecName: "config-data") pod "4189c7b4-f54f-4162-b97c-d210219b2784" (UID: "4189c7b4-f54f-4162-b97c-d210219b2784"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555096 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555190 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px89c\" (UniqueName: \"kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555263 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555357 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555455 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555470 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4189c7b4-f54f-4162-b97c-d210219b2784-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.555482 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gzgh\" (UniqueName: \"kubernetes.io/projected/4189c7b4-f54f-4162-b97c-d210219b2784-kube-api-access-7gzgh\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.657068 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.657161 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.657253 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.657283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px89c\" (UniqueName: \"kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.657901 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.663951 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.663990 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.676373 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px89c\" (UniqueName: \"kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c\") pod \"nova-api-0\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " pod="openstack/nova-api-0" Oct 02 11:18:10 crc kubenswrapper[4751]: I1002 11:18:10.733330 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.003211 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4189c7b4-f54f-4162-b97c-d210219b2784","Type":"ContainerDied","Data":"1e71085b1455931ec2231f575bc5dbc866c5c2d2ad2b9b51845ce10d9717ac05"} Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.003243 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.003275 4751 scope.go:117] "RemoveContainer" containerID="dea7e72f6536476024b63fdad6e73d0ec1ff03184e5f23d1ceb78efe16ec2046" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.038714 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.052635 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.061026 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.062521 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.064893 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.069872 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.166251 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.175259 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.175334 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpzhj\" (UniqueName: \"kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.175433 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.267565 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.268127 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.277528 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.277636 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpzhj\" (UniqueName: \"kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.277704 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.283202 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.283589 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.295932 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpzhj\" (UniqueName: \"kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj\") pod \"nova-scheduler-0\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.383289 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.562565 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4189c7b4-f54f-4162-b97c-d210219b2784" path="/var/lib/kubelet/pods/4189c7b4-f54f-4162-b97c-d210219b2784/volumes" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.563392 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca6623c-2f50-48ff-a82c-bc822c443f30" path="/var/lib/kubelet/pods/5ca6623c-2f50-48ff-a82c-bc822c443f30/volumes" Oct 02 11:18:11 crc kubenswrapper[4751]: I1002 11:18:11.818026 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.014019 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerStarted","Data":"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a"} Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.014063 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerStarted","Data":"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710"} Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.014077 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerStarted","Data":"5edb55d732f7d31a4d83ce3a2d46f1f41aef6736d235dab19c2d8f2f0185623f"} Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.016552 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1ea384e-64c0-49c7-8b5d-69bbed821f13","Type":"ContainerStarted","Data":"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4"} Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.016604 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1ea384e-64c0-49c7-8b5d-69bbed821f13","Type":"ContainerStarted","Data":"fae403dc6d1dc1e863bd90035c964fc0f8e31dfb2b8e71c63321aa9a97ae6bc8"} Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.031011 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.030994969 podStartE2EDuration="2.030994969s" podCreationTimestamp="2025-10-02 11:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:12.030130476 +0000 UTC m=+1574.084356926" watchObservedRunningTime="2025-10-02 11:18:12.030994969 +0000 UTC m=+1574.085221419" Oct 02 11:18:12 crc kubenswrapper[4751]: I1002 11:18:12.265259 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 11:18:13 crc kubenswrapper[4751]: I1002 11:18:13.043108 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.043085134 podStartE2EDuration="2.043085134s" podCreationTimestamp="2025-10-02 11:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:13.037662479 +0000 UTC m=+1575.091888929" watchObservedRunningTime="2025-10-02 11:18:13.043085134 +0000 UTC m=+1575.097311584" Oct 02 11:18:16 crc kubenswrapper[4751]: I1002 11:18:16.267833 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 11:18:16 crc kubenswrapper[4751]: I1002 11:18:16.268222 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 11:18:16 crc kubenswrapper[4751]: I1002 11:18:16.383693 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 11:18:17 crc kubenswrapper[4751]: I1002 11:18:17.281569 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:17 crc kubenswrapper[4751]: I1002 11:18:17.282274 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:20 crc kubenswrapper[4751]: I1002 11:18:20.734019 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:20 crc kubenswrapper[4751]: I1002 11:18:20.734671 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:21 crc kubenswrapper[4751]: I1002 11:18:21.384382 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 11:18:21 crc kubenswrapper[4751]: I1002 11:18:21.415284 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 11:18:21 crc kubenswrapper[4751]: I1002 11:18:21.815506 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:21 crc kubenswrapper[4751]: I1002 11:18:21.815448 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 11:18:22 crc kubenswrapper[4751]: I1002 11:18:22.155618 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 11:18:26 crc kubenswrapper[4751]: I1002 11:18:26.273560 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 11:18:26 crc kubenswrapper[4751]: I1002 11:18:26.275349 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 11:18:26 crc kubenswrapper[4751]: I1002 11:18:26.279908 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 11:18:27 crc kubenswrapper[4751]: I1002 11:18:27.162628 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 11:18:27 crc kubenswrapper[4751]: W1002 11:18:27.556551 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f.scope WatchSource:0}: Error finding container d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f: Status 404 returned error can't find the container with id d078a2aa5504172393929888f33fdc2b68e254dd0309840cb73f1af6b13ec39f Oct 02 11:18:27 crc kubenswrapper[4751]: W1002 11:18:27.557374 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0591cb81_3c5d_4c93_8a9c_88000aa3df5c.slice/crio-124bddfa11944597d4f7a60eb4433125f1170bea975e172a384ba3785ac596ec WatchSource:0}: Error finding container 124bddfa11944597d4f7a60eb4433125f1170bea975e172a384ba3785ac596ec: Status 404 returned error can't find the container with id 124bddfa11944597d4f7a60eb4433125f1170bea975e172a384ba3785ac596ec Oct 02 11:18:27 crc kubenswrapper[4751]: W1002 11:18:27.560509 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0591cb81_3c5d_4c93_8a9c_88000aa3df5c.slice/crio-758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad.scope WatchSource:0}: Error finding container 758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad: Status 404 returned error can't find the container with id 758e8e6c86e929c721d8c9693b46b60820e74cfeafe320c1d30a2a57cb35c9ad Oct 02 11:18:27 crc kubenswrapper[4751]: W1002 11:18:27.562608 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0591cb81_3c5d_4c93_8a9c_88000aa3df5c.slice/crio-fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1.scope WatchSource:0}: Error finding container fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1: Status 404 returned error can't find the container with id fa356ecae8e1295345251411eb37902cb97aa1a3f4d64cf5183b22ec3ccdaea1 Oct 02 11:18:27 crc kubenswrapper[4751]: E1002 11:18:27.865894 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc58d30ab_0099_48d9_a2be_15c630e1179d.slice/crio-3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-conmon-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc58d30ab_0099_48d9_a2be_15c630e1179d.slice/crio-conmon-3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4189c7b4_f54f_4162_b97c_d210219b2784.slice/crio-1e71085b1455931ec2231f575bc5dbc866c5c2d2ad2b9b51845ce10d9717ac05\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4189c7b4_f54f_4162_b97c_d210219b2784.slice\": RecentStats: unable to find data in memory cache]" Oct 02 11:18:27 crc kubenswrapper[4751]: I1002 11:18:27.944775 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.064989 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.107825 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data\") pod \"c58d30ab-0099-48d9-a2be-15c630e1179d\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.108009 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle\") pod \"c58d30ab-0099-48d9-a2be-15c630e1179d\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.108687 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sbz2\" (UniqueName: \"kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2\") pod \"c58d30ab-0099-48d9-a2be-15c630e1179d\" (UID: \"c58d30ab-0099-48d9-a2be-15c630e1179d\") " Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.113467 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2" (OuterVolumeSpecName: "kube-api-access-6sbz2") pod "c58d30ab-0099-48d9-a2be-15c630e1179d" (UID: "c58d30ab-0099-48d9-a2be-15c630e1179d"). InnerVolumeSpecName "kube-api-access-6sbz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.141469 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c58d30ab-0099-48d9-a2be-15c630e1179d" (UID: "c58d30ab-0099-48d9-a2be-15c630e1179d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.150855 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data" (OuterVolumeSpecName: "config-data") pod "c58d30ab-0099-48d9-a2be-15c630e1179d" (UID: "c58d30ab-0099-48d9-a2be-15c630e1179d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.172256 4751 generic.go:334] "Generic (PLEG): container finished" podID="c58d30ab-0099-48d9-a2be-15c630e1179d" containerID="3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3" exitCode=137 Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.172347 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c58d30ab-0099-48d9-a2be-15c630e1179d","Type":"ContainerDied","Data":"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3"} Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.172369 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.172403 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c58d30ab-0099-48d9-a2be-15c630e1179d","Type":"ContainerDied","Data":"00f3ee00c24c7ed24baee1ca9528b9c6d2bac981f6f2c767f56d76b6e1b536ca"} Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.172421 4751 scope.go:117] "RemoveContainer" containerID="3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.213253 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.213293 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58d30ab-0099-48d9-a2be-15c630e1179d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.213303 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sbz2\" (UniqueName: \"kubernetes.io/projected/c58d30ab-0099-48d9-a2be-15c630e1179d-kube-api-access-6sbz2\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.220217 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.223277 4751 scope.go:117] "RemoveContainer" containerID="3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3" Oct 02 11:18:28 crc kubenswrapper[4751]: E1002 11:18:28.226634 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3\": container with ID starting with 3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3 not found: ID does not exist" containerID="3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.226692 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3"} err="failed to get container status \"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3\": rpc error: code = NotFound desc = could not find container \"3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3\": container with ID starting with 3b79d3208bca7e39dc54dc07391051e0b2d2130687c14c7472868b196f253ef3 not found: ID does not exist" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.235106 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.260906 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:18:28 crc kubenswrapper[4751]: E1002 11:18:28.261676 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58d30ab-0099-48d9-a2be-15c630e1179d" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.261691 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58d30ab-0099-48d9-a2be-15c630e1179d" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.261870 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58d30ab-0099-48d9-a2be-15c630e1179d" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.262528 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.264397 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.265805 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.268230 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.278397 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.315479 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.315523 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.315573 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8scz\" (UniqueName: \"kubernetes.io/projected/9596583b-e70e-435f-b390-0f873a1da605-kube-api-access-r8scz\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.315606 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.315920 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.417378 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.417427 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.417507 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8scz\" (UniqueName: \"kubernetes.io/projected/9596583b-e70e-435f-b390-0f873a1da605-kube-api-access-r8scz\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.418184 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.418311 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.420817 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.420895 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.420951 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.423307 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9596583b-e70e-435f-b390-0f873a1da605-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.436100 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8scz\" (UniqueName: \"kubernetes.io/projected/9596583b-e70e-435f-b390-0f873a1da605-kube-api-access-r8scz\") pod \"nova-cell1-novncproxy-0\" (UID: \"9596583b-e70e-435f-b390-0f873a1da605\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:28 crc kubenswrapper[4751]: I1002 11:18:28.584956 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:29 crc kubenswrapper[4751]: I1002 11:18:29.134352 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 11:18:29 crc kubenswrapper[4751]: W1002 11:18:29.137398 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9596583b_e70e_435f_b390_0f873a1da605.slice/crio-ac9980e7a295021cc65b70a947e3af7a6111875ccf82b375def7cb775309d688 WatchSource:0}: Error finding container ac9980e7a295021cc65b70a947e3af7a6111875ccf82b375def7cb775309d688: Status 404 returned error can't find the container with id ac9980e7a295021cc65b70a947e3af7a6111875ccf82b375def7cb775309d688 Oct 02 11:18:29 crc kubenswrapper[4751]: I1002 11:18:29.187299 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9596583b-e70e-435f-b390-0f873a1da605","Type":"ContainerStarted","Data":"ac9980e7a295021cc65b70a947e3af7a6111875ccf82b375def7cb775309d688"} Oct 02 11:18:29 crc kubenswrapper[4751]: I1002 11:18:29.584284 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58d30ab-0099-48d9-a2be-15c630e1179d" path="/var/lib/kubelet/pods/c58d30ab-0099-48d9-a2be-15c630e1179d/volumes" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.199089 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9596583b-e70e-435f-b390-0f873a1da605","Type":"ContainerStarted","Data":"c510e5ce3c28661bab5c79971550001fa73843ac7693b099c2c034f8a9777019"} Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.219325 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.219308886 podStartE2EDuration="2.219308886s" podCreationTimestamp="2025-10-02 11:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:30.214668792 +0000 UTC m=+1592.268895242" watchObservedRunningTime="2025-10-02 11:18:30.219308886 +0000 UTC m=+1592.273535336" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.743739 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.744223 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.744618 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.744675 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.749572 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.750999 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.934613 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.936639 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:30 crc kubenswrapper[4751]: I1002 11:18:30.954450 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.076676 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.077674 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.077736 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.077835 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.078046 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx729\" (UniqueName: \"kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.078122 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180519 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180626 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180665 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180721 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180774 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx729\" (UniqueName: \"kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.180803 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.181675 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.181694 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.181823 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.181855 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.182445 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.206355 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx729\" (UniqueName: \"kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729\") pod \"dnsmasq-dns-5c7b6c5df9-fx7kv\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.268931 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.507023 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.507410 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:18:31 crc kubenswrapper[4751]: I1002 11:18:31.717834 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:18:31 crc kubenswrapper[4751]: W1002 11:18:31.719601 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7cade35_ef6d_4614_91b3_6fb1b010b31b.slice/crio-43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e WatchSource:0}: Error finding container 43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e: Status 404 returned error can't find the container with id 43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e Oct 02 11:18:32 crc kubenswrapper[4751]: I1002 11:18:32.220764 4751 generic.go:334] "Generic (PLEG): container finished" podID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerID="a0c97fd3714f56ccdd60f3d1fcd149aa9d49120281592d23f787660f554b7565" exitCode=0 Oct 02 11:18:32 crc kubenswrapper[4751]: I1002 11:18:32.222610 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" event={"ID":"e7cade35-ef6d-4614-91b3-6fb1b010b31b","Type":"ContainerDied","Data":"a0c97fd3714f56ccdd60f3d1fcd149aa9d49120281592d23f787660f554b7565"} Oct 02 11:18:32 crc kubenswrapper[4751]: I1002 11:18:32.222652 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" event={"ID":"e7cade35-ef6d-4614-91b3-6fb1b010b31b","Type":"ContainerStarted","Data":"43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e"} Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.018103 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.232008 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" event={"ID":"e7cade35-ef6d-4614-91b3-6fb1b010b31b","Type":"ContainerStarted","Data":"504b9d94d0a5bc36891f30d0cf058fff09bbea782297045d58a32769c1b9dd2f"} Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.233152 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-log" containerID="cri-o://d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710" gracePeriod=30 Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.233239 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-api" containerID="cri-o://3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a" gracePeriod=30 Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.267952 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" podStartSLOduration=3.267929601 podStartE2EDuration="3.267929601s" podCreationTimestamp="2025-10-02 11:18:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:33.256371412 +0000 UTC m=+1595.310597862" watchObservedRunningTime="2025-10-02 11:18:33.267929601 +0000 UTC m=+1595.322156051" Oct 02 11:18:33 crc kubenswrapper[4751]: I1002 11:18:33.586017 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.219348 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.219695 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-central-agent" containerID="cri-o://ce301b246e99c191f9083946459cef1b1fbe155015e8dbafca08ea162a31e232" gracePeriod=30 Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.219777 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-notification-agent" containerID="cri-o://e4995510d541c879a0992ef0e05126af02aead27d8708e9e0daba361c76661b4" gracePeriod=30 Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.219793 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="sg-core" containerID="cri-o://9a086cc41a2751992b4d40879db5a227aecde8b5e959356e7f30d81041832ceb" gracePeriod=30 Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.220022 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="proxy-httpd" containerID="cri-o://c67cd3ef1c4fe61ca5288a671379d49fa8971ba5c1aa63aebf51fa029ec17f8a" gracePeriod=30 Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.246705 4751 generic.go:334] "Generic (PLEG): container finished" podID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerID="d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710" exitCode=143 Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.246788 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerDied","Data":"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710"} Oct 02 11:18:34 crc kubenswrapper[4751]: I1002 11:18:34.247812 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.259777 4751 generic.go:334] "Generic (PLEG): container finished" podID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerID="c67cd3ef1c4fe61ca5288a671379d49fa8971ba5c1aa63aebf51fa029ec17f8a" exitCode=0 Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.260119 4751 generic.go:334] "Generic (PLEG): container finished" podID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerID="9a086cc41a2751992b4d40879db5a227aecde8b5e959356e7f30d81041832ceb" exitCode=2 Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.260133 4751 generic.go:334] "Generic (PLEG): container finished" podID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerID="ce301b246e99c191f9083946459cef1b1fbe155015e8dbafca08ea162a31e232" exitCode=0 Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.259820 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerDied","Data":"c67cd3ef1c4fe61ca5288a671379d49fa8971ba5c1aa63aebf51fa029ec17f8a"} Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.260200 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerDied","Data":"9a086cc41a2751992b4d40879db5a227aecde8b5e959356e7f30d81041832ceb"} Oct 02 11:18:35 crc kubenswrapper[4751]: I1002 11:18:35.260221 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerDied","Data":"ce301b246e99c191f9083946459cef1b1fbe155015e8dbafca08ea162a31e232"} Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.813241 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.987235 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data\") pod \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.987351 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle\") pod \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.987377 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs\") pod \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.987486 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px89c\" (UniqueName: \"kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c\") pod \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\" (UID: \"d86910ac-efcd-41ea-8e7a-5bf22f980a0f\") " Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.988120 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs" (OuterVolumeSpecName: "logs") pod "d86910ac-efcd-41ea-8e7a-5bf22f980a0f" (UID: "d86910ac-efcd-41ea-8e7a-5bf22f980a0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:36 crc kubenswrapper[4751]: I1002 11:18:36.992984 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c" (OuterVolumeSpecName: "kube-api-access-px89c") pod "d86910ac-efcd-41ea-8e7a-5bf22f980a0f" (UID: "d86910ac-efcd-41ea-8e7a-5bf22f980a0f"). InnerVolumeSpecName "kube-api-access-px89c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.031445 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data" (OuterVolumeSpecName: "config-data") pod "d86910ac-efcd-41ea-8e7a-5bf22f980a0f" (UID: "d86910ac-efcd-41ea-8e7a-5bf22f980a0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.033876 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d86910ac-efcd-41ea-8e7a-5bf22f980a0f" (UID: "d86910ac-efcd-41ea-8e7a-5bf22f980a0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.089429 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.089478 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.089494 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.089505 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px89c\" (UniqueName: \"kubernetes.io/projected/d86910ac-efcd-41ea-8e7a-5bf22f980a0f-kube-api-access-px89c\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.280505 4751 generic.go:334] "Generic (PLEG): container finished" podID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerID="3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a" exitCode=0 Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.280584 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.280582 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerDied","Data":"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a"} Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.281030 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d86910ac-efcd-41ea-8e7a-5bf22f980a0f","Type":"ContainerDied","Data":"5edb55d732f7d31a4d83ce3a2d46f1f41aef6736d235dab19c2d8f2f0185623f"} Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.281058 4751 scope.go:117] "RemoveContainer" containerID="3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.315687 4751 scope.go:117] "RemoveContainer" containerID="d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.316423 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.328866 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.340505 4751 scope.go:117] "RemoveContainer" containerID="3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a" Oct 02 11:18:37 crc kubenswrapper[4751]: E1002 11:18:37.340959 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a\": container with ID starting with 3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a not found: ID does not exist" containerID="3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.341000 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a"} err="failed to get container status \"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a\": rpc error: code = NotFound desc = could not find container \"3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a\": container with ID starting with 3d88abe5e5375d99224eb03e0be27b6971f0b072be8a56f16eaad43f18a6797a not found: ID does not exist" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.341029 4751 scope.go:117] "RemoveContainer" containerID="d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710" Oct 02 11:18:37 crc kubenswrapper[4751]: E1002 11:18:37.341524 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710\": container with ID starting with d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710 not found: ID does not exist" containerID="d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.341554 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710"} err="failed to get container status \"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710\": rpc error: code = NotFound desc = could not find container \"d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710\": container with ID starting with d1981a8b90160431be6b957ec08765951c9ee533c61930e5d1d6713b142e2710 not found: ID does not exist" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.346656 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:37 crc kubenswrapper[4751]: E1002 11:18:37.347195 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-api" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.347220 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-api" Oct 02 11:18:37 crc kubenswrapper[4751]: E1002 11:18:37.347250 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-log" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.347258 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-log" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.347502 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-log" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.347540 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" containerName="nova-api-api" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.348751 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.351634 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.351934 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.352105 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.354297 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.496128 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.497000 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.497192 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.497388 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wb8\" (UniqueName: \"kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.497533 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.497695 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.583573 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d86910ac-efcd-41ea-8e7a-5bf22f980a0f" path="/var/lib/kubelet/pods/d86910ac-efcd-41ea-8e7a-5bf22f980a0f/volumes" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.599835 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.599980 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68wb8\" (UniqueName: \"kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.600069 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.600149 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.600343 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.600427 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.600338 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.606053 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.606129 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.606484 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.606808 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.619300 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68wb8\" (UniqueName: \"kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8\") pod \"nova-api-0\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " pod="openstack/nova-api-0" Oct 02 11:18:37 crc kubenswrapper[4751]: I1002 11:18:37.674348 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:38 crc kubenswrapper[4751]: E1002 11:18:38.108024 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-conmon-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:18:38 crc kubenswrapper[4751]: I1002 11:18:38.128336 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:38 crc kubenswrapper[4751]: I1002 11:18:38.292740 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerStarted","Data":"65aa311474cc2f283f352a73426c72d2b894a61b7a655cfeabf65fbbf3da62e6"} Oct 02 11:18:38 crc kubenswrapper[4751]: I1002 11:18:38.585990 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:38 crc kubenswrapper[4751]: I1002 11:18:38.605416 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.305879 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerStarted","Data":"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14"} Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.306309 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerStarted","Data":"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd"} Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.309609 4751 generic.go:334] "Generic (PLEG): container finished" podID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerID="e4995510d541c879a0992ef0e05126af02aead27d8708e9e0daba361c76661b4" exitCode=0 Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.309732 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerDied","Data":"e4995510d541c879a0992ef0e05126af02aead27d8708e9e0daba361c76661b4"} Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.309779 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bfe3610c-0efc-4358-9abc-5c476a8e7851","Type":"ContainerDied","Data":"96003922d481185750396f4310672af285bc506acdb843cad29320f894c63b96"} Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.309795 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96003922d481185750396f4310672af285bc506acdb843cad29320f894c63b96" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.329692 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.346164 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.346138453 podStartE2EDuration="2.346138453s" podCreationTimestamp="2025-10-02 11:18:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:39.326758325 +0000 UTC m=+1601.380984775" watchObservedRunningTime="2025-10-02 11:18:39.346138453 +0000 UTC m=+1601.400364903" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.350679 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.534265 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.534704 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.534865 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9s6c\" (UniqueName: \"kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.534904 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.535063 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.535110 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.535153 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.535236 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle\") pod \"bfe3610c-0efc-4358-9abc-5c476a8e7851\" (UID: \"bfe3610c-0efc-4358-9abc-5c476a8e7851\") " Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.536867 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.537167 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.546500 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts" (OuterVolumeSpecName: "scripts") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.548746 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c" (OuterVolumeSpecName: "kube-api-access-j9s6c") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "kube-api-access-j9s6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.593469 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.619238 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637708 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9s6c\" (UniqueName: \"kubernetes.io/projected/bfe3610c-0efc-4358-9abc-5c476a8e7851-kube-api-access-j9s6c\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637744 4751 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637754 4751 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bfe3610c-0efc-4358-9abc-5c476a8e7851-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637765 4751 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637773 4751 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.637782 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.644246 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.675911 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data" (OuterVolumeSpecName: "config-data") pod "bfe3610c-0efc-4358-9abc-5c476a8e7851" (UID: "bfe3610c-0efc-4358-9abc-5c476a8e7851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.740423 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.740538 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe3610c-0efc-4358-9abc-5c476a8e7851-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.755632 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rtkfm"] Oct 02 11:18:39 crc kubenswrapper[4751]: E1002 11:18:39.756063 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="proxy-httpd" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756085 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="proxy-httpd" Oct 02 11:18:39 crc kubenswrapper[4751]: E1002 11:18:39.756117 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-central-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756127 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-central-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: E1002 11:18:39.756162 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-notification-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756207 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-notification-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: E1002 11:18:39.756236 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="sg-core" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756244 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="sg-core" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756461 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-notification-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756493 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="ceilometer-central-agent" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756516 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="proxy-httpd" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.756532 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" containerName="sg-core" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.757198 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rtkfm"] Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.757321 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.760251 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.762573 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.944365 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snn8n\" (UniqueName: \"kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.944420 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.944989 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:39 crc kubenswrapper[4751]: I1002 11:18:39.945243 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.047370 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snn8n\" (UniqueName: \"kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.047418 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.047470 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.047522 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.053034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.055663 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.056523 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.063053 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snn8n\" (UniqueName: \"kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n\") pod \"nova-cell1-cell-mapping-rtkfm\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.081093 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.318292 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.356368 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.364444 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.382403 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.385251 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.386980 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.388395 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.388611 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.391279 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562003 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562048 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtvs\" (UniqueName: \"kubernetes.io/projected/9a0ea287-4160-468d-b606-efdbc47e9c50-kube-api-access-qgtvs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562378 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-scripts\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562534 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-run-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562719 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.562916 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.563046 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-config-data\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.563104 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-log-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.664915 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.664995 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665020 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-config-data\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-log-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665114 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665132 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtvs\" (UniqueName: \"kubernetes.io/projected/9a0ea287-4160-468d-b606-efdbc47e9c50-kube-api-access-qgtvs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665164 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-scripts\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665211 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-run-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.665669 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-run-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.666683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a0ea287-4160-468d-b606-efdbc47e9c50-log-httpd\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.672271 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-scripts\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.672908 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.673857 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.682886 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.683654 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0ea287-4160-468d-b606-efdbc47e9c50-config-data\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.689215 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtvs\" (UniqueName: \"kubernetes.io/projected/9a0ea287-4160-468d-b606-efdbc47e9c50-kube-api-access-qgtvs\") pod \"ceilometer-0\" (UID: \"9a0ea287-4160-468d-b606-efdbc47e9c50\") " pod="openstack/ceilometer-0" Oct 02 11:18:40 crc kubenswrapper[4751]: I1002 11:18:40.703402 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.135282 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rtkfm"] Oct 02 11:18:41 crc kubenswrapper[4751]: W1002 11:18:41.183808 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a0ea287_4160_468d_b606_efdbc47e9c50.slice/crio-9938d679c746dafbe49b6b986932bd733cbb5aaca97995e59cb242c100721bc2 WatchSource:0}: Error finding container 9938d679c746dafbe49b6b986932bd733cbb5aaca97995e59cb242c100721bc2: Status 404 returned error can't find the container with id 9938d679c746dafbe49b6b986932bd733cbb5aaca97995e59cb242c100721bc2 Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.184072 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.270437 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.342503 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rtkfm" event={"ID":"f387d75d-a7dc-4ca1-8a55-d24502420fbb","Type":"ContainerStarted","Data":"d961c7bdb06f272c351e6bc2065dc3bd357371cebf470fa61108f7b01d736514"} Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.342627 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rtkfm" event={"ID":"f387d75d-a7dc-4ca1-8a55-d24502420fbb","Type":"ContainerStarted","Data":"73d8010eac1e4dc9ae59dabfc239836b9bbeff0e576a54bc3363ea9f44c9db98"} Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.343929 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"9938d679c746dafbe49b6b986932bd733cbb5aaca97995e59cb242c100721bc2"} Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.356381 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.356660 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="dnsmasq-dns" containerID="cri-o://d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1" gracePeriod=10 Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.364438 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rtkfm" podStartSLOduration=2.364421895 podStartE2EDuration="2.364421895s" podCreationTimestamp="2025-10-02 11:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:41.363905951 +0000 UTC m=+1603.418132401" watchObservedRunningTime="2025-10-02 11:18:41.364421895 +0000 UTC m=+1603.418648345" Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.562064 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe3610c-0efc-4358-9abc-5c476a8e7851" path="/var/lib/kubelet/pods/bfe3610c-0efc-4358-9abc-5c476a8e7851/volumes" Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.880342 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991302 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5bbs\" (UniqueName: \"kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991411 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991463 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991519 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991544 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.991579 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb\") pod \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\" (UID: \"42567d4c-6c29-4c0c-b5d8-3c9c63f08326\") " Oct 02 11:18:41 crc kubenswrapper[4751]: I1002 11:18:41.996501 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs" (OuterVolumeSpecName: "kube-api-access-k5bbs") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "kube-api-access-k5bbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.046896 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.050765 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.051723 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config" (OuterVolumeSpecName: "config") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.057032 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.074637 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "42567d4c-6c29-4c0c-b5d8-3c9c63f08326" (UID: "42567d4c-6c29-4c0c-b5d8-3c9c63f08326"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094192 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5bbs\" (UniqueName: \"kubernetes.io/projected/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-kube-api-access-k5bbs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094228 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094243 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094257 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094268 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.094278 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/42567d4c-6c29-4c0c-b5d8-3c9c63f08326-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.354605 4751 generic.go:334] "Generic (PLEG): container finished" podID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerID="d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1" exitCode=0 Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.354674 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" event={"ID":"42567d4c-6c29-4c0c-b5d8-3c9c63f08326","Type":"ContainerDied","Data":"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1"} Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.354701 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" event={"ID":"42567d4c-6c29-4c0c-b5d8-3c9c63f08326","Type":"ContainerDied","Data":"2d44cc2cb65ae424ff0d5d714d665073bd5bcdb431d5e2f30c225278e691021d"} Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.354701 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-2ll65" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.354716 4751 scope.go:117] "RemoveContainer" containerID="d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.356103 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"b921d098eaef98400b81e4dfcb4b97fac188a0f675c78762ab537dd6a3f9014a"} Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.374978 4751 scope.go:117] "RemoveContainer" containerID="4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.403392 4751 scope.go:117] "RemoveContainer" containerID="d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1" Oct 02 11:18:42 crc kubenswrapper[4751]: E1002 11:18:42.404608 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1\": container with ID starting with d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1 not found: ID does not exist" containerID="d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.404641 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1"} err="failed to get container status \"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1\": rpc error: code = NotFound desc = could not find container \"d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1\": container with ID starting with d972e8855e3b19a9896429937444105ef79becef3da028f71ab72dd49ed536f1 not found: ID does not exist" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.404662 4751 scope.go:117] "RemoveContainer" containerID="4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968" Oct 02 11:18:42 crc kubenswrapper[4751]: E1002 11:18:42.404879 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968\": container with ID starting with 4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968 not found: ID does not exist" containerID="4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.404907 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968"} err="failed to get container status \"4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968\": rpc error: code = NotFound desc = could not find container \"4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968\": container with ID starting with 4e56d4bcbd0d3d496fff79b804fe57547148bcdec770e2bc9a7a97c6440d8968 not found: ID does not exist" Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.405698 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:18:42 crc kubenswrapper[4751]: I1002 11:18:42.415466 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-2ll65"] Oct 02 11:18:43 crc kubenswrapper[4751]: I1002 11:18:43.369635 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"3cf200985a3c509bb544c7cbef7fa19f907982e8a91664e40f44658c02396d39"} Oct 02 11:18:43 crc kubenswrapper[4751]: I1002 11:18:43.370000 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"8bd03c8350da92c55c058eb9a70006a3bb14c0c2050af578987d13c46db0d36c"} Oct 02 11:18:43 crc kubenswrapper[4751]: I1002 11:18:43.563253 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" path="/var/lib/kubelet/pods/42567d4c-6c29-4c0c-b5d8-3c9c63f08326/volumes" Oct 02 11:18:45 crc kubenswrapper[4751]: I1002 11:18:45.389568 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"32584b74c16f59fdf684289d393ac12907e2804d825d23a1db437e47f0a9182f"} Oct 02 11:18:45 crc kubenswrapper[4751]: I1002 11:18:45.390347 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 11:18:45 crc kubenswrapper[4751]: I1002 11:18:45.419982 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.663209511 podStartE2EDuration="5.419949576s" podCreationTimestamp="2025-10-02 11:18:40 +0000 UTC" firstStartedPulling="2025-10-02 11:18:41.187027412 +0000 UTC m=+1603.241253862" lastFinishedPulling="2025-10-02 11:18:44.943767477 +0000 UTC m=+1606.997993927" observedRunningTime="2025-10-02 11:18:45.410228676 +0000 UTC m=+1607.464455146" watchObservedRunningTime="2025-10-02 11:18:45.419949576 +0000 UTC m=+1607.474176046" Oct 02 11:18:46 crc kubenswrapper[4751]: I1002 11:18:46.406781 4751 generic.go:334] "Generic (PLEG): container finished" podID="f387d75d-a7dc-4ca1-8a55-d24502420fbb" containerID="d961c7bdb06f272c351e6bc2065dc3bd357371cebf470fa61108f7b01d736514" exitCode=0 Oct 02 11:18:46 crc kubenswrapper[4751]: I1002 11:18:46.407008 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rtkfm" event={"ID":"f387d75d-a7dc-4ca1-8a55-d24502420fbb","Type":"ContainerDied","Data":"d961c7bdb06f272c351e6bc2065dc3bd357371cebf470fa61108f7b01d736514"} Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.675431 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.675788 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.801865 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.915518 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snn8n\" (UniqueName: \"kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n\") pod \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.915668 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle\") pod \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.915787 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data\") pod \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.915952 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts\") pod \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\" (UID: \"f387d75d-a7dc-4ca1-8a55-d24502420fbb\") " Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.921753 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts" (OuterVolumeSpecName: "scripts") pod "f387d75d-a7dc-4ca1-8a55-d24502420fbb" (UID: "f387d75d-a7dc-4ca1-8a55-d24502420fbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.924346 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n" (OuterVolumeSpecName: "kube-api-access-snn8n") pod "f387d75d-a7dc-4ca1-8a55-d24502420fbb" (UID: "f387d75d-a7dc-4ca1-8a55-d24502420fbb"). InnerVolumeSpecName "kube-api-access-snn8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.948485 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f387d75d-a7dc-4ca1-8a55-d24502420fbb" (UID: "f387d75d-a7dc-4ca1-8a55-d24502420fbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:47 crc kubenswrapper[4751]: I1002 11:18:47.956980 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data" (OuterVolumeSpecName: "config-data") pod "f387d75d-a7dc-4ca1-8a55-d24502420fbb" (UID: "f387d75d-a7dc-4ca1-8a55-d24502420fbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.018416 4751 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.018463 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snn8n\" (UniqueName: \"kubernetes.io/projected/f387d75d-a7dc-4ca1-8a55-d24502420fbb-kube-api-access-snn8n\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.018489 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.018513 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f387d75d-a7dc-4ca1-8a55-d24502420fbb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:48 crc kubenswrapper[4751]: E1002 11:18:48.363985 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-conmon-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.428566 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rtkfm" event={"ID":"f387d75d-a7dc-4ca1-8a55-d24502420fbb","Type":"ContainerDied","Data":"73d8010eac1e4dc9ae59dabfc239836b9bbeff0e576a54bc3363ea9f44c9db98"} Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.428616 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73d8010eac1e4dc9ae59dabfc239836b9bbeff0e576a54bc3363ea9f44c9db98" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.428682 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rtkfm" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.617095 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.623612 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-log" containerID="cri-o://9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14" gracePeriod=30 Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.624123 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-api" containerID="cri-o://a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd" gracePeriod=30 Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.634233 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.634401 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.639230 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.639527 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" containerName="nova-scheduler-scheduler" containerID="cri-o://6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4" gracePeriod=30 Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.655273 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.655479 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" containerID="cri-o://ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206" gracePeriod=30 Oct 02 11:18:48 crc kubenswrapper[4751]: I1002 11:18:48.655810 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" containerID="cri-o://e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4" gracePeriod=30 Oct 02 11:18:49 crc kubenswrapper[4751]: I1002 11:18:49.438143 4751 generic.go:334] "Generic (PLEG): container finished" podID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerID="ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206" exitCode=143 Oct 02 11:18:49 crc kubenswrapper[4751]: I1002 11:18:49.438335 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerDied","Data":"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206"} Oct 02 11:18:49 crc kubenswrapper[4751]: I1002 11:18:49.441947 4751 generic.go:334] "Generic (PLEG): container finished" podID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerID="9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14" exitCode=143 Oct 02 11:18:49 crc kubenswrapper[4751]: I1002 11:18:49.441979 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerDied","Data":"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14"} Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.259148 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.372656 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle\") pod \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.372795 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpzhj\" (UniqueName: \"kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj\") pod \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.372911 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data\") pod \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\" (UID: \"d1ea384e-64c0-49c7-8b5d-69bbed821f13\") " Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.379457 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj" (OuterVolumeSpecName: "kube-api-access-qpzhj") pod "d1ea384e-64c0-49c7-8b5d-69bbed821f13" (UID: "d1ea384e-64c0-49c7-8b5d-69bbed821f13"). InnerVolumeSpecName "kube-api-access-qpzhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.405517 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1ea384e-64c0-49c7-8b5d-69bbed821f13" (UID: "d1ea384e-64c0-49c7-8b5d-69bbed821f13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.422405 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data" (OuterVolumeSpecName: "config-data") pod "d1ea384e-64c0-49c7-8b5d-69bbed821f13" (UID: "d1ea384e-64c0-49c7-8b5d-69bbed821f13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.453533 4751 generic.go:334] "Generic (PLEG): container finished" podID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" containerID="6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4" exitCode=0 Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.453588 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1ea384e-64c0-49c7-8b5d-69bbed821f13","Type":"ContainerDied","Data":"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4"} Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.453623 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d1ea384e-64c0-49c7-8b5d-69bbed821f13","Type":"ContainerDied","Data":"fae403dc6d1dc1e863bd90035c964fc0f8e31dfb2b8e71c63321aa9a97ae6bc8"} Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.453642 4751 scope.go:117] "RemoveContainer" containerID="6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.453801 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.476880 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.476904 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ea384e-64c0-49c7-8b5d-69bbed821f13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.476916 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpzhj\" (UniqueName: \"kubernetes.io/projected/d1ea384e-64c0-49c7-8b5d-69bbed821f13-kube-api-access-qpzhj\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.501304 4751 scope.go:117] "RemoveContainer" containerID="6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4" Oct 02 11:18:50 crc kubenswrapper[4751]: E1002 11:18:50.502313 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4\": container with ID starting with 6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4 not found: ID does not exist" containerID="6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.502364 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4"} err="failed to get container status \"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4\": rpc error: code = NotFound desc = could not find container \"6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4\": container with ID starting with 6ac4a7413aa8a505faa61d1a2a3a2a3aec9f3d2cf72a9286f2194e285776d1d4 not found: ID does not exist" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.515066 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.536842 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.551291 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:50 crc kubenswrapper[4751]: E1002 11:18:50.551716 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f387d75d-a7dc-4ca1-8a55-d24502420fbb" containerName="nova-manage" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.551735 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f387d75d-a7dc-4ca1-8a55-d24502420fbb" containerName="nova-manage" Oct 02 11:18:50 crc kubenswrapper[4751]: E1002 11:18:50.551860 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="dnsmasq-dns" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.551868 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="dnsmasq-dns" Oct 02 11:18:50 crc kubenswrapper[4751]: E1002 11:18:50.551884 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" containerName="nova-scheduler-scheduler" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.551890 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" containerName="nova-scheduler-scheduler" Oct 02 11:18:50 crc kubenswrapper[4751]: E1002 11:18:50.551897 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="init" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.551903 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="init" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.552100 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" containerName="nova-scheduler-scheduler" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.552128 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="42567d4c-6c29-4c0c-b5d8-3c9c63f08326" containerName="dnsmasq-dns" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.552160 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f387d75d-a7dc-4ca1-8a55-d24502420fbb" containerName="nova-manage" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.553037 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.556159 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.597970 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.693605 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-config-data\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.693726 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.694024 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l59f6\" (UniqueName: \"kubernetes.io/projected/3fb989ac-8357-4e3c-af91-5378e0a874a9-kube-api-access-l59f6\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.795192 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-config-data\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.795511 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.795618 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l59f6\" (UniqueName: \"kubernetes.io/projected/3fb989ac-8357-4e3c-af91-5378e0a874a9-kube-api-access-l59f6\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.799677 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-config-data\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.801262 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fb989ac-8357-4e3c-af91-5378e0a874a9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.812557 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l59f6\" (UniqueName: \"kubernetes.io/projected/3fb989ac-8357-4e3c-af91-5378e0a874a9-kube-api-access-l59f6\") pod \"nova-scheduler-0\" (UID: \"3fb989ac-8357-4e3c-af91-5378e0a874a9\") " pod="openstack/nova-scheduler-0" Oct 02 11:18:50 crc kubenswrapper[4751]: I1002 11:18:50.884420 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 11:18:51 crc kubenswrapper[4751]: I1002 11:18:51.329717 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 11:18:51 crc kubenswrapper[4751]: W1002 11:18:51.330436 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fb989ac_8357_4e3c_af91_5378e0a874a9.slice/crio-1b9c386a7921d29159c71e02decb5675fd11a8febf4433bfcfa68e3ee6f72557 WatchSource:0}: Error finding container 1b9c386a7921d29159c71e02decb5675fd11a8febf4433bfcfa68e3ee6f72557: Status 404 returned error can't find the container with id 1b9c386a7921d29159c71e02decb5675fd11a8febf4433bfcfa68e3ee6f72557 Oct 02 11:18:51 crc kubenswrapper[4751]: I1002 11:18:51.464111 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3fb989ac-8357-4e3c-af91-5378e0a874a9","Type":"ContainerStarted","Data":"1b9c386a7921d29159c71e02decb5675fd11a8febf4433bfcfa68e3ee6f72557"} Oct 02 11:18:51 crc kubenswrapper[4751]: I1002 11:18:51.564357 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ea384e-64c0-49c7-8b5d-69bbed821f13" path="/var/lib/kubelet/pods/d1ea384e-64c0-49c7-8b5d-69bbed821f13/volumes" Oct 02 11:18:51 crc kubenswrapper[4751]: I1002 11:18:51.787393 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:43654->10.217.0.195:8775: read: connection reset by peer" Oct 02 11:18:51 crc kubenswrapper[4751]: I1002 11:18:51.787753 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:43638->10.217.0.195:8775: read: connection reset by peer" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.260353 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.422966 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4ssr\" (UniqueName: \"kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr\") pod \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.423113 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle\") pod \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.423206 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs\") pod \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.423307 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data\") pod \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.423359 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs\") pod \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\" (UID: \"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63\") " Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.424619 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs" (OuterVolumeSpecName: "logs") pod "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" (UID: "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.428422 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr" (OuterVolumeSpecName: "kube-api-access-z4ssr") pod "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" (UID: "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63"). InnerVolumeSpecName "kube-api-access-z4ssr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.458401 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data" (OuterVolumeSpecName: "config-data") pod "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" (UID: "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.463618 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" (UID: "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.483137 4751 generic.go:334] "Generic (PLEG): container finished" podID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerID="e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4" exitCode=0 Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.483205 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.483222 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerDied","Data":"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4"} Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.485878 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"58c4ecb7-093d-4c0f-85a2-c1f019bf0c63","Type":"ContainerDied","Data":"19e2dec392aa9e70eac6f5aa099170e99b5f55cac7bfb8e726652e9c54cabc80"} Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.485929 4751 scope.go:117] "RemoveContainer" containerID="e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.486410 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3fb989ac-8357-4e3c-af91-5378e0a874a9","Type":"ContainerStarted","Data":"da13bb87b97f83ccb88d846e7e77db8538319beaac1ed11c768bbef255895710"} Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.495750 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" (UID: "58c4ecb7-093d-4c0f-85a2-c1f019bf0c63"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.508830 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.508814094 podStartE2EDuration="2.508814094s" podCreationTimestamp="2025-10-02 11:18:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:52.506275306 +0000 UTC m=+1614.560501776" watchObservedRunningTime="2025-10-02 11:18:52.508814094 +0000 UTC m=+1614.563040534" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.525901 4751 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.525950 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4ssr\" (UniqueName: \"kubernetes.io/projected/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-kube-api-access-z4ssr\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.525966 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.525977 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.525988 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.559615 4751 scope.go:117] "RemoveContainer" containerID="ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.580841 4751 scope.go:117] "RemoveContainer" containerID="e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4" Oct 02 11:18:52 crc kubenswrapper[4751]: E1002 11:18:52.581452 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4\": container with ID starting with e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4 not found: ID does not exist" containerID="e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.581515 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4"} err="failed to get container status \"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4\": rpc error: code = NotFound desc = could not find container \"e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4\": container with ID starting with e6b33292ebac9f877c110d86d7abadd35196d14c45ff7ced2bc82573c9f065b4 not found: ID does not exist" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.581555 4751 scope.go:117] "RemoveContainer" containerID="ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206" Oct 02 11:18:52 crc kubenswrapper[4751]: E1002 11:18:52.582046 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206\": container with ID starting with ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206 not found: ID does not exist" containerID="ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.582074 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206"} err="failed to get container status \"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206\": rpc error: code = NotFound desc = could not find container \"ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206\": container with ID starting with ad706a1fcdbe3d781b5c3caf3b433b98804e09806e0279d516e4ccbbf8c7e206 not found: ID does not exist" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.830383 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.853682 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.867266 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:52 crc kubenswrapper[4751]: E1002 11:18:52.867689 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.867708 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" Oct 02 11:18:52 crc kubenswrapper[4751]: E1002 11:18:52.867739 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.867746 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.867916 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-log" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.867939 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" containerName="nova-metadata-metadata" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.870074 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.872230 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.876681 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.888883 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.933668 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-config-data\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.933776 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.933835 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-logs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.933926 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk4mx\" (UniqueName: \"kubernetes.io/projected/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-kube-api-access-vk4mx\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:52 crc kubenswrapper[4751]: I1002 11:18:52.933974 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.035825 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.036132 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-logs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.036350 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk4mx\" (UniqueName: \"kubernetes.io/projected/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-kube-api-access-vk4mx\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.036486 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.036701 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-config-data\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.037193 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-logs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.040804 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.041602 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.042717 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-config-data\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.057099 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk4mx\" (UniqueName: \"kubernetes.io/projected/1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2-kube-api-access-vk4mx\") pod \"nova-metadata-0\" (UID: \"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2\") " pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.191795 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.569217 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c4ecb7-093d-4c0f-85a2-c1f019bf0c63" path="/var/lib/kubelet/pods/58c4ecb7-093d-4c0f-85a2-c1f019bf0c63/volumes" Oct 02 11:18:53 crc kubenswrapper[4751]: I1002 11:18:53.629416 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.481956 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.511905 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2","Type":"ContainerStarted","Data":"73eaa4ec85a07e16f4f18bed01a14eea49746a1ad8a08208ac4944c68a0d10b2"} Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.511944 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2","Type":"ContainerStarted","Data":"a1d14299181bbd65ab36bf62aeae6ee9265ba20ec5f9d027baf55bbbb7b5d170"} Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.511952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2","Type":"ContainerStarted","Data":"ea499afd7296a91bb11c7129307c2f0e8c7f4721039fceb9f6aebc18ac1ed621"} Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.514640 4751 generic.go:334] "Generic (PLEG): container finished" podID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerID="a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd" exitCode=0 Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.514686 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.514688 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerDied","Data":"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd"} Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.514746 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"109ee027-189c-4f98-8b20-455d65e7d6cd","Type":"ContainerDied","Data":"65aa311474cc2f283f352a73426c72d2b894a61b7a655cfeabf65fbbf3da62e6"} Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.514768 4751 scope.go:117] "RemoveContainer" containerID="a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.536474 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.536456246 podStartE2EDuration="2.536456246s" podCreationTimestamp="2025-10-02 11:18:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:54.534772251 +0000 UTC m=+1616.588998721" watchObservedRunningTime="2025-10-02 11:18:54.536456246 +0000 UTC m=+1616.590682696" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.541641 4751 scope.go:117] "RemoveContainer" containerID="9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.561491 4751 scope.go:117] "RemoveContainer" containerID="a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd" Oct 02 11:18:54 crc kubenswrapper[4751]: E1002 11:18:54.562090 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd\": container with ID starting with a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd not found: ID does not exist" containerID="a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.562123 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd"} err="failed to get container status \"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd\": rpc error: code = NotFound desc = could not find container \"a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd\": container with ID starting with a545f4a6aeab2c08ea258202e1868555521fca7c54a94ea2ab2dc08a373a52cd not found: ID does not exist" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.562146 4751 scope.go:117] "RemoveContainer" containerID="9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14" Oct 02 11:18:54 crc kubenswrapper[4751]: E1002 11:18:54.562553 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14\": container with ID starting with 9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14 not found: ID does not exist" containerID="9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.562573 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14"} err="failed to get container status \"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14\": rpc error: code = NotFound desc = could not find container \"9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14\": container with ID starting with 9cede8161d78de99743704901db98eea71d8c627de0199a1890a05eb87fd0d14 not found: ID does not exist" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669606 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669743 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669815 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669838 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669910 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.669964 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68wb8\" (UniqueName: \"kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8\") pod \"109ee027-189c-4f98-8b20-455d65e7d6cd\" (UID: \"109ee027-189c-4f98-8b20-455d65e7d6cd\") " Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.671122 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs" (OuterVolumeSpecName: "logs") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.676438 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8" (OuterVolumeSpecName: "kube-api-access-68wb8") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "kube-api-access-68wb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.696496 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data" (OuterVolumeSpecName: "config-data") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.704129 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.730352 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.732682 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "109ee027-189c-4f98-8b20-455d65e7d6cd" (UID: "109ee027-189c-4f98-8b20-455d65e7d6cd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772525 4751 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772587 4751 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/109ee027-189c-4f98-8b20-455d65e7d6cd-logs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772606 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772624 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68wb8\" (UniqueName: \"kubernetes.io/projected/109ee027-189c-4f98-8b20-455d65e7d6cd-kube-api-access-68wb8\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772649 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.772667 4751 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/109ee027-189c-4f98-8b20-455d65e7d6cd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.862109 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.875883 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.890358 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:54 crc kubenswrapper[4751]: E1002 11:18:54.890913 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-log" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.890941 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-log" Oct 02 11:18:54 crc kubenswrapper[4751]: E1002 11:18:54.890968 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-api" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.890978 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-api" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.891236 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-api" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.891276 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" containerName="nova-api-log" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.893919 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.896678 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.897955 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.898035 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.899110 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.976599 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-internal-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.976914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-config-data\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.977110 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-logs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.977474 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvfsh\" (UniqueName: \"kubernetes.io/projected/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-kube-api-access-lvfsh\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.977708 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-public-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:54 crc kubenswrapper[4751]: I1002 11:18:54.977901 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079298 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvfsh\" (UniqueName: \"kubernetes.io/projected/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-kube-api-access-lvfsh\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079593 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-public-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079643 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079676 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-internal-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079694 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-config-data\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.079725 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-logs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.080113 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-logs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.084644 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-config-data\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.084683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-public-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.085127 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-internal-tls-certs\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.085324 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.105844 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvfsh\" (UniqueName: \"kubernetes.io/projected/609f5b43-0fb5-4044-95a2-ff1f0ab991ea-kube-api-access-lvfsh\") pod \"nova-api-0\" (UID: \"609f5b43-0fb5-4044-95a2-ff1f0ab991ea\") " pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.272225 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.563986 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109ee027-189c-4f98-8b20-455d65e7d6cd" path="/var/lib/kubelet/pods/109ee027-189c-4f98-8b20-455d65e7d6cd/volumes" Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.705062 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 11:18:55 crc kubenswrapper[4751]: W1002 11:18:55.705801 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod609f5b43_0fb5_4044_95a2_ff1f0ab991ea.slice/crio-231a45cdb2e657c0aa773a26ff87d4bd96783032bfcd86cd90c17afee565aa65 WatchSource:0}: Error finding container 231a45cdb2e657c0aa773a26ff87d4bd96783032bfcd86cd90c17afee565aa65: Status 404 returned error can't find the container with id 231a45cdb2e657c0aa773a26ff87d4bd96783032bfcd86cd90c17afee565aa65 Oct 02 11:18:55 crc kubenswrapper[4751]: I1002 11:18:55.885426 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 11:18:56 crc kubenswrapper[4751]: I1002 11:18:56.539972 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"609f5b43-0fb5-4044-95a2-ff1f0ab991ea","Type":"ContainerStarted","Data":"c93b75f285ff5c500a2491de9410d1c12f04e165b3dfdddcdfa764c4239f9057"} Oct 02 11:18:56 crc kubenswrapper[4751]: I1002 11:18:56.540433 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"609f5b43-0fb5-4044-95a2-ff1f0ab991ea","Type":"ContainerStarted","Data":"d88945580b34ab7f588dfa9e652837a7b963967636c39ebe8e0ed28d01182309"} Oct 02 11:18:56 crc kubenswrapper[4751]: I1002 11:18:56.540513 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"609f5b43-0fb5-4044-95a2-ff1f0ab991ea","Type":"ContainerStarted","Data":"231a45cdb2e657c0aa773a26ff87d4bd96783032bfcd86cd90c17afee565aa65"} Oct 02 11:18:56 crc kubenswrapper[4751]: I1002 11:18:56.572466 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.57243016 podStartE2EDuration="2.57243016s" podCreationTimestamp="2025-10-02 11:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:18:56.561266829 +0000 UTC m=+1618.615493289" watchObservedRunningTime="2025-10-02 11:18:56.57243016 +0000 UTC m=+1618.626656640" Oct 02 11:18:58 crc kubenswrapper[4751]: I1002 11:18:58.192160 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:58 crc kubenswrapper[4751]: I1002 11:18:58.192484 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 11:18:58 crc kubenswrapper[4751]: E1002 11:18:58.646418 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a42e63_2379_4384_9bea_f756390a1935.slice/crio-conmon-548e57bb47656b170b628fb019bd2d210bf1fb9f4ba166d9b42ac3c96a5fbec7.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:19:00 crc kubenswrapper[4751]: I1002 11:19:00.885392 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 11:19:00 crc kubenswrapper[4751]: I1002 11:19:00.915655 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.507116 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.507193 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.507243 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.508015 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.508080 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" gracePeriod=600 Oct 02 11:19:01 crc kubenswrapper[4751]: I1002 11:19:01.611452 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 11:19:01 crc kubenswrapper[4751]: E1002 11:19:01.642605 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:02 crc kubenswrapper[4751]: I1002 11:19:02.590892 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" exitCode=0 Oct 02 11:19:02 crc kubenswrapper[4751]: I1002 11:19:02.590952 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168"} Oct 02 11:19:02 crc kubenswrapper[4751]: I1002 11:19:02.591040 4751 scope.go:117] "RemoveContainer" containerID="2726ac25611ed7b2d41edbecdcf3a5ff54a9543f383fe1766e67db5fe1de94d9" Oct 02 11:19:02 crc kubenswrapper[4751]: I1002 11:19:02.591834 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:19:02 crc kubenswrapper[4751]: E1002 11:19:02.592148 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:03 crc kubenswrapper[4751]: I1002 11:19:03.193393 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 11:19:03 crc kubenswrapper[4751]: I1002 11:19:03.193783 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 11:19:04 crc kubenswrapper[4751]: I1002 11:19:04.209413 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:19:04 crc kubenswrapper[4751]: I1002 11:19:04.209424 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:19:05 crc kubenswrapper[4751]: I1002 11:19:05.273269 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:19:05 crc kubenswrapper[4751]: I1002 11:19:05.273535 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 11:19:06 crc kubenswrapper[4751]: I1002 11:19:06.291435 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="609f5b43-0fb5-4044-95a2-ff1f0ab991ea" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:19:06 crc kubenswrapper[4751]: I1002 11:19:06.291835 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="609f5b43-0fb5-4044-95a2-ff1f0ab991ea" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 11:19:10 crc kubenswrapper[4751]: I1002 11:19:10.712903 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 11:19:13 crc kubenswrapper[4751]: I1002 11:19:13.197971 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 11:19:13 crc kubenswrapper[4751]: I1002 11:19:13.198593 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 11:19:13 crc kubenswrapper[4751]: I1002 11:19:13.202587 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 11:19:13 crc kubenswrapper[4751]: I1002 11:19:13.550213 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:19:13 crc kubenswrapper[4751]: E1002 11:19:13.550451 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:13 crc kubenswrapper[4751]: I1002 11:19:13.698279 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.281914 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.282643 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.284033 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.289324 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.713686 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 11:19:15 crc kubenswrapper[4751]: I1002 11:19:15.721759 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 11:19:24 crc kubenswrapper[4751]: I1002 11:19:24.989948 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:25 crc kubenswrapper[4751]: I1002 11:19:25.550482 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:19:25 crc kubenswrapper[4751]: E1002 11:19:25.550784 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:26 crc kubenswrapper[4751]: I1002 11:19:26.637573 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:29 crc kubenswrapper[4751]: I1002 11:19:29.486085 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="rabbitmq" containerID="cri-o://bc6a75489d3fb857d4be16da3b9fd49d8b5d3dfd494165122057d9571f8923a4" gracePeriod=604796 Oct 02 11:19:30 crc kubenswrapper[4751]: I1002 11:19:30.504833 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="rabbitmq" containerID="cri-o://91fe3fd870d8dddb3232de4e25185991bcc93aa0b82d75d6afe4d2c5efceafb2" gracePeriod=604797 Oct 02 11:19:35 crc kubenswrapper[4751]: I1002 11:19:35.904291 4751 generic.go:334] "Generic (PLEG): container finished" podID="433c0079-6192-4515-8c6d-5f035612f44b" containerID="bc6a75489d3fb857d4be16da3b9fd49d8b5d3dfd494165122057d9571f8923a4" exitCode=0 Oct 02 11:19:35 crc kubenswrapper[4751]: I1002 11:19:35.904407 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerDied","Data":"bc6a75489d3fb857d4be16da3b9fd49d8b5d3dfd494165122057d9571f8923a4"} Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.085425 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.183978 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqr24\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184089 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184139 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184226 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184268 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184302 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184367 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184407 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184436 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184465 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.184500 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins\") pod \"433c0079-6192-4515-8c6d-5f035612f44b\" (UID: \"433c0079-6192-4515-8c6d-5f035612f44b\") " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.185956 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.188453 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.188783 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.193590 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.193602 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24" (OuterVolumeSpecName: "kube-api-access-vqr24") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "kube-api-access-vqr24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.195912 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info" (OuterVolumeSpecName: "pod-info") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.200243 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.202068 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.226447 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data" (OuterVolumeSpecName: "config-data") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.253990 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf" (OuterVolumeSpecName: "server-conf") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287234 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287281 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqr24\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-kube-api-access-vqr24\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287298 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287311 4751 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/433c0079-6192-4515-8c6d-5f035612f44b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287358 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287372 4751 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/433c0079-6192-4515-8c6d-5f035612f44b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287384 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287394 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287404 4751 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.287415 4751 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/433c0079-6192-4515-8c6d-5f035612f44b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.306882 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "433c0079-6192-4515-8c6d-5f035612f44b" (UID: "433c0079-6192-4515-8c6d-5f035612f44b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.320636 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.390057 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/433c0079-6192-4515-8c6d-5f035612f44b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.390114 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.741459 4751 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.924563 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"433c0079-6192-4515-8c6d-5f035612f44b","Type":"ContainerDied","Data":"ec900eeada206aa078b73b92d7513119d67c69fccfcd33f677eaccdf7351f171"} Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.924651 4751 scope.go:117] "RemoveContainer" containerID="bc6a75489d3fb857d4be16da3b9fd49d8b5d3dfd494165122057d9571f8923a4" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.924594 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.930274 4751 generic.go:334] "Generic (PLEG): container finished" podID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerID="91fe3fd870d8dddb3232de4e25185991bcc93aa0b82d75d6afe4d2c5efceafb2" exitCode=0 Oct 02 11:19:36 crc kubenswrapper[4751]: I1002 11:19:36.930328 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerDied","Data":"91fe3fd870d8dddb3232de4e25185991bcc93aa0b82d75d6afe4d2c5efceafb2"} Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.072720 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.075156 4751 scope.go:117] "RemoveContainer" containerID="8d0714368474f6ee386d468f8266d4e2d0c5105a27266788a8998b264c68bb24" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.086587 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.096478 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.145079 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:37 crc kubenswrapper[4751]: E1002 11:19:37.145773 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="setup-container" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.145794 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="setup-container" Oct 02 11:19:37 crc kubenswrapper[4751]: E1002 11:19:37.145812 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.145820 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: E1002 11:19:37.145862 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.145869 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: E1002 11:19:37.145886 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="setup-container" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.145892 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="setup-container" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.146139 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.146160 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="433c0079-6192-4515-8c6d-5f035612f44b" containerName="rabbitmq" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.147825 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.150787 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.150889 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.150922 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.150943 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.151097 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.151368 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4q2c4" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.153627 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.160152 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205403 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205468 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205516 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205540 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx9f8\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205584 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205636 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205699 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205745 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205769 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205802 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.205822 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf\") pod \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\" (UID: \"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b\") " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.206623 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.207740 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.210510 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.211048 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.224257 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.224325 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8" (OuterVolumeSpecName: "kube-api-access-tx9f8") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "kube-api-access-tx9f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.224333 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info" (OuterVolumeSpecName: "pod-info") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.224362 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.243561 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data" (OuterVolumeSpecName: "config-data") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.275682 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf" (OuterVolumeSpecName: "server-conf") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308483 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90870cce-dc95-4298-8220-a9e0e61f315a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308558 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308593 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308623 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw425\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-kube-api-access-bw425\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308660 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308757 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308789 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308910 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308965 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-config-data\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.308996 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90870cce-dc95-4298-8220-a9e0e61f315a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309033 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309128 4751 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309149 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309162 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309253 4751 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309266 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309277 4751 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309288 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx9f8\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-kube-api-access-tx9f8\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309299 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309309 4751 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.309319 4751 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.330025 4751 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.331875 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" (UID: "341995a5-cf9c-42ce-9afb-b7b50b0c6c4b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412028 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412088 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412116 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412141 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-config-data\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412187 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90870cce-dc95-4298-8220-a9e0e61f315a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412310 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90870cce-dc95-4298-8220-a9e0e61f315a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412349 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412375 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412398 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw425\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-kube-api-access-bw425\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412431 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412525 4751 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412542 4751 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412659 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.412783 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.413555 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.413986 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-config-data\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.414055 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.414307 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/90870cce-dc95-4298-8220-a9e0e61f315a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.415864 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/90870cce-dc95-4298-8220-a9e0e61f315a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.415917 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.416389 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/90870cce-dc95-4298-8220-a9e0e61f315a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.418621 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.428885 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw425\" (UniqueName: \"kubernetes.io/projected/90870cce-dc95-4298-8220-a9e0e61f315a-kube-api-access-bw425\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.454458 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"90870cce-dc95-4298-8220-a9e0e61f315a\") " pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.472277 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.551574 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:19:37 crc kubenswrapper[4751]: E1002 11:19:37.551905 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.564330 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="433c0079-6192-4515-8c6d-5f035612f44b" path="/var/lib/kubelet/pods/433c0079-6192-4515-8c6d-5f035612f44b/volumes" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.937398 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.946364 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"341995a5-cf9c-42ce-9afb-b7b50b0c6c4b","Type":"ContainerDied","Data":"b39c67bf08faca1f6ff89d0963058cd6fe8375995163a674a7d9dcb3abf282e9"} Oct 02 11:19:37 crc kubenswrapper[4751]: W1002 11:19:37.946384 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90870cce_dc95_4298_8220_a9e0e61f315a.slice/crio-72d22551100f02e21d06de6b38be89fbcf5242928ac72e06cfba3b9972362acf WatchSource:0}: Error finding container 72d22551100f02e21d06de6b38be89fbcf5242928ac72e06cfba3b9972362acf: Status 404 returned error can't find the container with id 72d22551100f02e21d06de6b38be89fbcf5242928ac72e06cfba3b9972362acf Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.946409 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:37 crc kubenswrapper[4751]: I1002 11:19:37.946438 4751 scope.go:117] "RemoveContainer" containerID="91fe3fd870d8dddb3232de4e25185991bcc93aa0b82d75d6afe4d2c5efceafb2" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.085785 4751 scope.go:117] "RemoveContainer" containerID="3432c82b9e0622ec5f5ece1b81a44963f17f80a50ec2661d0b17a2e1ba4c5a3a" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.121294 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.130722 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.156036 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.158156 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.160393 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.160420 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.160725 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.161045 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.161100 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.162029 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zv4vl" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.162451 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.220487 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227299 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227384 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227462 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227525 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227617 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.227656 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b4c8\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-kube-api-access-6b4c8\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.228368 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.228541 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34d3499-dc03-46d9-8dde-43c0f06c659b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.228685 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.228901 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34d3499-dc03-46d9-8dde-43c0f06c659b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.228943 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.330574 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.330919 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b4c8\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-kube-api-access-6b4c8\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.330980 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34d3499-dc03-46d9-8dde-43c0f06c659b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331041 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331127 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34d3499-dc03-46d9-8dde-43c0f06c659b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331153 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331222 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331253 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331291 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331321 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331794 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.331917 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.332246 4751 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.332350 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.332468 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.332671 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d34d3499-dc03-46d9-8dde-43c0f06c659b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.335496 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.335529 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d34d3499-dc03-46d9-8dde-43c0f06c659b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.336192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d34d3499-dc03-46d9-8dde-43c0f06c659b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.337135 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.351414 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b4c8\" (UniqueName: \"kubernetes.io/projected/d34d3499-dc03-46d9-8dde-43c0f06c659b-kube-api-access-6b4c8\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.366237 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d34d3499-dc03-46d9-8dde-43c0f06c659b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.495515 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.958209 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90870cce-dc95-4298-8220-a9e0e61f315a","Type":"ContainerStarted","Data":"72d22551100f02e21d06de6b38be89fbcf5242928ac72e06cfba3b9972362acf"} Oct 02 11:19:38 crc kubenswrapper[4751]: I1002 11:19:38.974747 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 11:19:39 crc kubenswrapper[4751]: W1002 11:19:39.060880 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd34d3499_dc03_46d9_8dde_43c0f06c659b.slice/crio-5b4820b39520f8e54e8236291c212693352969ea366c12c8f769807ce1af2b0c WatchSource:0}: Error finding container 5b4820b39520f8e54e8236291c212693352969ea366c12c8f769807ce1af2b0c: Status 404 returned error can't find the container with id 5b4820b39520f8e54e8236291c212693352969ea366c12c8f769807ce1af2b0c Oct 02 11:19:39 crc kubenswrapper[4751]: I1002 11:19:39.565156 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341995a5-cf9c-42ce-9afb-b7b50b0c6c4b" path="/var/lib/kubelet/pods/341995a5-cf9c-42ce-9afb-b7b50b0c6c4b/volumes" Oct 02 11:19:39 crc kubenswrapper[4751]: I1002 11:19:39.975379 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34d3499-dc03-46d9-8dde-43c0f06c659b","Type":"ContainerStarted","Data":"5b4820b39520f8e54e8236291c212693352969ea366c12c8f769807ce1af2b0c"} Oct 02 11:19:39 crc kubenswrapper[4751]: I1002 11:19:39.977902 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90870cce-dc95-4298-8220-a9e0e61f315a","Type":"ContainerStarted","Data":"b0116a048ff425da11927486faf579e4234ef45a6fa5dfb58ea7717142a18025"} Oct 02 11:19:41 crc kubenswrapper[4751]: I1002 11:19:41.000395 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34d3499-dc03-46d9-8dde-43c0f06c659b","Type":"ContainerStarted","Data":"960615326ae718664cae214bedeab2a79a8ac10fded5399eae634c95de77b40c"} Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.366555 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.369997 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.373796 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.400811 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.423939 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424002 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rcn7\" (UniqueName: \"kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424036 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424215 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424301 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.424508 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526185 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526267 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rcn7\" (UniqueName: \"kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526296 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526342 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526378 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.526487 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.527480 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.527733 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.527753 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.528064 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.528413 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.530217 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.561142 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rcn7\" (UniqueName: \"kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7\") pod \"dnsmasq-dns-5576978c7c-8npxf\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:42 crc kubenswrapper[4751]: I1002 11:19:42.690922 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:43 crc kubenswrapper[4751]: W1002 11:19:43.136734 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b3138d_84ba_4a09_af02_71aab4bafef7.slice/crio-d686ea3c5f373e15cb077e84e8329a5dcfc30792620627d0af4d5a75e306659d WatchSource:0}: Error finding container d686ea3c5f373e15cb077e84e8329a5dcfc30792620627d0af4d5a75e306659d: Status 404 returned error can't find the container with id d686ea3c5f373e15cb077e84e8329a5dcfc30792620627d0af4d5a75e306659d Oct 02 11:19:43 crc kubenswrapper[4751]: I1002 11:19:43.145998 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:19:44 crc kubenswrapper[4751]: I1002 11:19:44.030138 4751 generic.go:334] "Generic (PLEG): container finished" podID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerID="48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8" exitCode=0 Oct 02 11:19:44 crc kubenswrapper[4751]: I1002 11:19:44.030206 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" event={"ID":"30b3138d-84ba-4a09-af02-71aab4bafef7","Type":"ContainerDied","Data":"48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8"} Oct 02 11:19:44 crc kubenswrapper[4751]: I1002 11:19:44.030454 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" event={"ID":"30b3138d-84ba-4a09-af02-71aab4bafef7","Type":"ContainerStarted","Data":"d686ea3c5f373e15cb077e84e8329a5dcfc30792620627d0af4d5a75e306659d"} Oct 02 11:19:45 crc kubenswrapper[4751]: I1002 11:19:45.041291 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" event={"ID":"30b3138d-84ba-4a09-af02-71aab4bafef7","Type":"ContainerStarted","Data":"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148"} Oct 02 11:19:45 crc kubenswrapper[4751]: I1002 11:19:45.042045 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:45 crc kubenswrapper[4751]: I1002 11:19:45.060984 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" podStartSLOduration=3.060939922 podStartE2EDuration="3.060939922s" podCreationTimestamp="2025-10-02 11:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:19:45.058752603 +0000 UTC m=+1667.112979073" watchObservedRunningTime="2025-10-02 11:19:45.060939922 +0000 UTC m=+1667.115166402" Oct 02 11:19:50 crc kubenswrapper[4751]: I1002 11:19:50.549946 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:19:50 crc kubenswrapper[4751]: E1002 11:19:50.550935 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.692344 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.747157 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.747469 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="dnsmasq-dns" containerID="cri-o://504b9d94d0a5bc36891f30d0cf058fff09bbea782297045d58a32769c1b9dd2f" gracePeriod=10 Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.938013 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-slxrn"] Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.941371 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:52 crc kubenswrapper[4751]: I1002 11:19:52.956609 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-slxrn"] Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033253 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033326 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033371 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033407 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-config\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033461 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033489 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4tw6\" (UniqueName: \"kubernetes.io/projected/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-kube-api-access-d4tw6\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.033534 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.138919 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139078 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139112 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139156 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139215 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-config\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139271 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.139307 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4tw6\" (UniqueName: \"kubernetes.io/projected/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-kube-api-access-d4tw6\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.141198 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.143978 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.144533 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-config\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.144781 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.145160 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.145689 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.163662 4751 generic.go:334] "Generic (PLEG): container finished" podID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerID="504b9d94d0a5bc36891f30d0cf058fff09bbea782297045d58a32769c1b9dd2f" exitCode=0 Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.163718 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" event={"ID":"e7cade35-ef6d-4614-91b3-6fb1b010b31b","Type":"ContainerDied","Data":"504b9d94d0a5bc36891f30d0cf058fff09bbea782297045d58a32769c1b9dd2f"} Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.163752 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" event={"ID":"e7cade35-ef6d-4614-91b3-6fb1b010b31b","Type":"ContainerDied","Data":"43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e"} Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.163766 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f6069b597d139637b9884c4aca8fc9cc4d03eef340e7160d906a49993d417e" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.167750 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4tw6\" (UniqueName: \"kubernetes.io/projected/af3ed200-3eef-4cc4-8b65-a0b92d3a8b45-kube-api-access-d4tw6\") pod \"dnsmasq-dns-8c6f6df99-slxrn\" (UID: \"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45\") " pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.220785 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.260224 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.344660 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.344990 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.345476 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.345512 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.345586 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.345685 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx729\" (UniqueName: \"kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729\") pod \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\" (UID: \"e7cade35-ef6d-4614-91b3-6fb1b010b31b\") " Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.354163 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729" (OuterVolumeSpecName: "kube-api-access-wx729") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "kube-api-access-wx729". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.404779 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.406527 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config" (OuterVolumeSpecName: "config") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.416154 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.419688 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.420871 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e7cade35-ef6d-4614-91b3-6fb1b010b31b" (UID: "e7cade35-ef6d-4614-91b3-6fb1b010b31b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447896 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx729\" (UniqueName: \"kubernetes.io/projected/e7cade35-ef6d-4614-91b3-6fb1b010b31b-kube-api-access-wx729\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447932 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447947 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447958 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447971 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.447983 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7cade35-ef6d-4614-91b3-6fb1b010b31b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:19:53 crc kubenswrapper[4751]: I1002 11:19:53.700490 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-slxrn"] Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.173124 4751 generic.go:334] "Generic (PLEG): container finished" podID="af3ed200-3eef-4cc4-8b65-a0b92d3a8b45" containerID="c3491d0f22b51d81d36bd601e743f1d4f31e27881308f1f4b0ec245f864a583e" exitCode=0 Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.173180 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" event={"ID":"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45","Type":"ContainerDied","Data":"c3491d0f22b51d81d36bd601e743f1d4f31e27881308f1f4b0ec245f864a583e"} Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.173540 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-fx7kv" Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.173536 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" event={"ID":"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45","Type":"ContainerStarted","Data":"0b7105f14f70c52d535d9e96f592f2946c597e84cb74a6cdbe62f0f1ddb996a5"} Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.215563 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:19:54 crc kubenswrapper[4751]: I1002 11:19:54.228678 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-fx7kv"] Oct 02 11:19:55 crc kubenswrapper[4751]: I1002 11:19:55.185795 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" event={"ID":"af3ed200-3eef-4cc4-8b65-a0b92d3a8b45","Type":"ContainerStarted","Data":"4f5e42a960e3183f8eee94390173dcb334ee85297bd21a58d75476f0f9242833"} Oct 02 11:19:55 crc kubenswrapper[4751]: I1002 11:19:55.186377 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:19:55 crc kubenswrapper[4751]: I1002 11:19:55.216747 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" podStartSLOduration=3.216729909 podStartE2EDuration="3.216729909s" podCreationTimestamp="2025-10-02 11:19:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:19:55.203327587 +0000 UTC m=+1677.257554097" watchObservedRunningTime="2025-10-02 11:19:55.216729909 +0000 UTC m=+1677.270956359" Oct 02 11:19:55 crc kubenswrapper[4751]: I1002 11:19:55.565461 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" path="/var/lib/kubelet/pods/e7cade35-ef6d-4614-91b3-6fb1b010b31b/volumes" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.262560 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-slxrn" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.324948 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.325302 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="dnsmasq-dns" containerID="cri-o://6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148" gracePeriod=10 Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.773356 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846537 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846720 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846747 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846866 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846911 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846936 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rcn7\" (UniqueName: \"kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.846971 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb\") pod \"30b3138d-84ba-4a09-af02-71aab4bafef7\" (UID: \"30b3138d-84ba-4a09-af02-71aab4bafef7\") " Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.853437 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7" (OuterVolumeSpecName: "kube-api-access-7rcn7") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "kube-api-access-7rcn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.895314 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.896297 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config" (OuterVolumeSpecName: "config") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.897297 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.898879 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.901943 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.905445 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30b3138d-84ba-4a09-af02-71aab4bafef7" (UID: "30b3138d-84ba-4a09-af02-71aab4bafef7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949629 4751 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-config\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949671 4751 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949682 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rcn7\" (UniqueName: \"kubernetes.io/projected/30b3138d-84ba-4a09-af02-71aab4bafef7-kube-api-access-7rcn7\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949690 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949700 4751 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949713 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:03 crc kubenswrapper[4751]: I1002 11:20:03.949724 4751 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b3138d-84ba-4a09-af02-71aab4bafef7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.281338 4751 generic.go:334] "Generic (PLEG): container finished" podID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerID="6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148" exitCode=0 Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.281387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" event={"ID":"30b3138d-84ba-4a09-af02-71aab4bafef7","Type":"ContainerDied","Data":"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148"} Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.281440 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.281747 4751 scope.go:117] "RemoveContainer" containerID="6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.281725 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-8npxf" event={"ID":"30b3138d-84ba-4a09-af02-71aab4bafef7","Type":"ContainerDied","Data":"d686ea3c5f373e15cb077e84e8329a5dcfc30792620627d0af4d5a75e306659d"} Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.309476 4751 scope.go:117] "RemoveContainer" containerID="48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.314895 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.325075 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-8npxf"] Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.338401 4751 scope.go:117] "RemoveContainer" containerID="6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148" Oct 02 11:20:04 crc kubenswrapper[4751]: E1002 11:20:04.339643 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148\": container with ID starting with 6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148 not found: ID does not exist" containerID="6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.339688 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148"} err="failed to get container status \"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148\": rpc error: code = NotFound desc = could not find container \"6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148\": container with ID starting with 6317e8314d94e8e103552b8eaee91115c2758b2ceccf0dd75cb0710cbf306148 not found: ID does not exist" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.339717 4751 scope.go:117] "RemoveContainer" containerID="48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8" Oct 02 11:20:04 crc kubenswrapper[4751]: E1002 11:20:04.340021 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8\": container with ID starting with 48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8 not found: ID does not exist" containerID="48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.340050 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8"} err="failed to get container status \"48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8\": rpc error: code = NotFound desc = could not find container \"48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8\": container with ID starting with 48e13c60e03f7dc86166e2b316346c6a5692aa801a125166ac53a3aad92009a8 not found: ID does not exist" Oct 02 11:20:04 crc kubenswrapper[4751]: I1002 11:20:04.549830 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:20:04 crc kubenswrapper[4751]: E1002 11:20:04.550109 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:20:05 crc kubenswrapper[4751]: I1002 11:20:05.571541 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" path="/var/lib/kubelet/pods/30b3138d-84ba-4a09-af02-71aab4bafef7/volumes" Oct 02 11:20:08 crc kubenswrapper[4751]: I1002 11:20:08.381851 4751 scope.go:117] "RemoveContainer" containerID="c4b1ba3fc0d08e4a2c9125ebe704a9e6704ff9971f808dc396e3a11ea62832b8" Oct 02 11:20:08 crc kubenswrapper[4751]: I1002 11:20:08.406881 4751 scope.go:117] "RemoveContainer" containerID="68a263fe7a00e1b5527ce62c4038d2a1285a89c7b1c52046ab5cf03d8c97af0b" Oct 02 11:20:12 crc kubenswrapper[4751]: I1002 11:20:12.358462 4751 generic.go:334] "Generic (PLEG): container finished" podID="90870cce-dc95-4298-8220-a9e0e61f315a" containerID="b0116a048ff425da11927486faf579e4234ef45a6fa5dfb58ea7717142a18025" exitCode=0 Oct 02 11:20:12 crc kubenswrapper[4751]: I1002 11:20:12.358561 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90870cce-dc95-4298-8220-a9e0e61f315a","Type":"ContainerDied","Data":"b0116a048ff425da11927486faf579e4234ef45a6fa5dfb58ea7717142a18025"} Oct 02 11:20:13 crc kubenswrapper[4751]: I1002 11:20:13.370586 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"90870cce-dc95-4298-8220-a9e0e61f315a","Type":"ContainerStarted","Data":"e7ce06aff93948779c38bcbaae4a1e8c49f9a868d71898bba92ca15d146fd7cc"} Oct 02 11:20:13 crc kubenswrapper[4751]: I1002 11:20:13.371252 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 11:20:13 crc kubenswrapper[4751]: I1002 11:20:13.401252 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.401227839 podStartE2EDuration="36.401227839s" podCreationTimestamp="2025-10-02 11:19:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:20:13.396915042 +0000 UTC m=+1695.451141492" watchObservedRunningTime="2025-10-02 11:20:13.401227839 +0000 UTC m=+1695.455454319" Oct 02 11:20:14 crc kubenswrapper[4751]: I1002 11:20:14.385237 4751 generic.go:334] "Generic (PLEG): container finished" podID="d34d3499-dc03-46d9-8dde-43c0f06c659b" containerID="960615326ae718664cae214bedeab2a79a8ac10fded5399eae634c95de77b40c" exitCode=0 Oct 02 11:20:14 crc kubenswrapper[4751]: I1002 11:20:14.385470 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34d3499-dc03-46d9-8dde-43c0f06c659b","Type":"ContainerDied","Data":"960615326ae718664cae214bedeab2a79a8ac10fded5399eae634c95de77b40c"} Oct 02 11:20:15 crc kubenswrapper[4751]: I1002 11:20:15.395405 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d34d3499-dc03-46d9-8dde-43c0f06c659b","Type":"ContainerStarted","Data":"f528a3bc68c03cbff0491b6e1e62b3f60c96793129de63de2a12cf73459f425a"} Oct 02 11:20:15 crc kubenswrapper[4751]: I1002 11:20:15.396625 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:20:15 crc kubenswrapper[4751]: I1002 11:20:15.421851 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.421830823 podStartE2EDuration="37.421830823s" podCreationTimestamp="2025-10-02 11:19:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:20:15.415454501 +0000 UTC m=+1697.469680961" watchObservedRunningTime="2025-10-02 11:20:15.421830823 +0000 UTC m=+1697.476057263" Oct 02 11:20:17 crc kubenswrapper[4751]: I1002 11:20:17.550336 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:20:17 crc kubenswrapper[4751]: E1002 11:20:17.550931 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.341681 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb"] Oct 02 11:20:20 crc kubenswrapper[4751]: E1002 11:20:20.342515 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342530 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: E1002 11:20:20.342542 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342550 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: E1002 11:20:20.342579 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="init" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342588 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="init" Oct 02 11:20:20 crc kubenswrapper[4751]: E1002 11:20:20.342598 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="init" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342606 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="init" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342861 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7cade35-ef6d-4614-91b3-6fb1b010b31b" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.342884 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b3138d-84ba-4a09-af02-71aab4bafef7" containerName="dnsmasq-dns" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.343917 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.345970 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.345986 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.346292 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.346423 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.356853 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb"] Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.462688 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.463275 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.463743 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwzcd\" (UniqueName: \"kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.463853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.567316 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.567436 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.567662 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.567806 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwzcd\" (UniqueName: \"kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.576099 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.576930 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.583575 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.587534 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwzcd\" (UniqueName: \"kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:20 crc kubenswrapper[4751]: I1002 11:20:20.665588 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:20:21 crc kubenswrapper[4751]: I1002 11:20:21.222351 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb"] Oct 02 11:20:21 crc kubenswrapper[4751]: W1002 11:20:21.225236 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7928fef2_025b_4bab_ab81_839abc70715b.slice/crio-9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2 WatchSource:0}: Error finding container 9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2: Status 404 returned error can't find the container with id 9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2 Oct 02 11:20:21 crc kubenswrapper[4751]: I1002 11:20:21.447451 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" event={"ID":"7928fef2-025b-4bab-ab81-839abc70715b","Type":"ContainerStarted","Data":"9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2"} Oct 02 11:20:27 crc kubenswrapper[4751]: I1002 11:20:27.475389 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 11:20:28 crc kubenswrapper[4751]: I1002 11:20:28.500345 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 11:20:32 crc kubenswrapper[4751]: I1002 11:20:32.555588 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:20:32 crc kubenswrapper[4751]: E1002 11:20:32.556965 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:20:38 crc kubenswrapper[4751]: E1002 11:20:38.759026 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:926e8bf36b3e6fc31124fa53557dcd5e01fdedc7" Oct 02 11:20:38 crc kubenswrapper[4751]: E1002 11:20:38.759607 4751 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:926e8bf36b3e6fc31124fa53557dcd5e01fdedc7" Oct 02 11:20:38 crc kubenswrapper[4751]: E1002 11:20:38.759762 4751 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 02 11:20:38 crc kubenswrapper[4751]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:926e8bf36b3e6fc31124fa53557dcd5e01fdedc7,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 02 11:20:38 crc kubenswrapper[4751]: - hosts: all Oct 02 11:20:38 crc kubenswrapper[4751]: strategy: linear Oct 02 11:20:38 crc kubenswrapper[4751]: tasks: Oct 02 11:20:38 crc kubenswrapper[4751]: - name: Enable podified-repos Oct 02 11:20:38 crc kubenswrapper[4751]: become: true Oct 02 11:20:38 crc kubenswrapper[4751]: ansible.builtin.shell: | Oct 02 11:20:38 crc kubenswrapper[4751]: set -euxo pipefail Oct 02 11:20:38 crc kubenswrapper[4751]: pushd /var/tmp Oct 02 11:20:38 crc kubenswrapper[4751]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 02 11:20:38 crc kubenswrapper[4751]: pushd repo-setup-main Oct 02 11:20:38 crc kubenswrapper[4751]: python3 -m venv ./venv Oct 02 11:20:38 crc kubenswrapper[4751]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 02 11:20:38 crc kubenswrapper[4751]: ./venv/bin/repo-setup current-podified -b antelope Oct 02 11:20:38 crc kubenswrapper[4751]: popd Oct 02 11:20:38 crc kubenswrapper[4751]: rm -rf repo-setup-main Oct 02 11:20:38 crc kubenswrapper[4751]: Oct 02 11:20:38 crc kubenswrapper[4751]: Oct 02 11:20:38 crc kubenswrapper[4751]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 02 11:20:38 crc kubenswrapper[4751]: edpm_override_hosts: openstack-edpm-ipam Oct 02 11:20:38 crc kubenswrapper[4751]: edpm_service_type: repo-setup Oct 02 11:20:38 crc kubenswrapper[4751]: Oct 02 11:20:38 crc kubenswrapper[4751]: Oct 02 11:20:38 crc kubenswrapper[4751]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwzcd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb_openstack(7928fef2-025b-4bab-ab81-839abc70715b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 02 11:20:38 crc kubenswrapper[4751]: > logger="UnhandledError" Oct 02 11:20:38 crc kubenswrapper[4751]: E1002 11:20:38.760997 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" podUID="7928fef2-025b-4bab-ab81-839abc70715b" Oct 02 11:20:39 crc kubenswrapper[4751]: E1002 11:20:39.647249 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/openstack-k8s-operators/openstack-ansibleee-runner:926e8bf36b3e6fc31124fa53557dcd5e01fdedc7\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" podUID="7928fef2-025b-4bab-ab81-839abc70715b" Oct 02 11:20:47 crc kubenswrapper[4751]: I1002 11:20:47.550727 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:20:47 crc kubenswrapper[4751]: E1002 11:20:47.551743 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:20:55 crc kubenswrapper[4751]: I1002 11:20:55.826455 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" event={"ID":"7928fef2-025b-4bab-ab81-839abc70715b","Type":"ContainerStarted","Data":"d56dc6d64e217a623ebb5f790b6fb185d99f834b558644aa15cde36c6addb15b"} Oct 02 11:20:55 crc kubenswrapper[4751]: I1002 11:20:55.854068 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" podStartSLOduration=2.318596405 podStartE2EDuration="35.854044755s" podCreationTimestamp="2025-10-02 11:20:20 +0000 UTC" firstStartedPulling="2025-10-02 11:20:21.228059829 +0000 UTC m=+1703.282286289" lastFinishedPulling="2025-10-02 11:20:54.763508169 +0000 UTC m=+1736.817734639" observedRunningTime="2025-10-02 11:20:55.849107812 +0000 UTC m=+1737.903334272" watchObservedRunningTime="2025-10-02 11:20:55.854044755 +0000 UTC m=+1737.908271215" Oct 02 11:21:02 crc kubenswrapper[4751]: I1002 11:21:02.550701 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:21:02 crc kubenswrapper[4751]: E1002 11:21:02.552524 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:21:07 crc kubenswrapper[4751]: I1002 11:21:07.953786 4751 generic.go:334] "Generic (PLEG): container finished" podID="7928fef2-025b-4bab-ab81-839abc70715b" containerID="d56dc6d64e217a623ebb5f790b6fb185d99f834b558644aa15cde36c6addb15b" exitCode=0 Oct 02 11:21:07 crc kubenswrapper[4751]: I1002 11:21:07.953935 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" event={"ID":"7928fef2-025b-4bab-ab81-839abc70715b","Type":"ContainerDied","Data":"d56dc6d64e217a623ebb5f790b6fb185d99f834b558644aa15cde36c6addb15b"} Oct 02 11:21:08 crc kubenswrapper[4751]: I1002 11:21:08.535531 4751 scope.go:117] "RemoveContainer" containerID="42a720db4d767a81d1a1d892ec3f9effbda006ede1bf2910a2d7e2914333993c" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.369896 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.523918 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle\") pod \"7928fef2-025b-4bab-ab81-839abc70715b\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.524138 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key\") pod \"7928fef2-025b-4bab-ab81-839abc70715b\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.524191 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory\") pod \"7928fef2-025b-4bab-ab81-839abc70715b\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.524960 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwzcd\" (UniqueName: \"kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd\") pod \"7928fef2-025b-4bab-ab81-839abc70715b\" (UID: \"7928fef2-025b-4bab-ab81-839abc70715b\") " Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.530364 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd" (OuterVolumeSpecName: "kube-api-access-hwzcd") pod "7928fef2-025b-4bab-ab81-839abc70715b" (UID: "7928fef2-025b-4bab-ab81-839abc70715b"). InnerVolumeSpecName "kube-api-access-hwzcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.531070 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7928fef2-025b-4bab-ab81-839abc70715b" (UID: "7928fef2-025b-4bab-ab81-839abc70715b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.559345 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7928fef2-025b-4bab-ab81-839abc70715b" (UID: "7928fef2-025b-4bab-ab81-839abc70715b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.575842 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory" (OuterVolumeSpecName: "inventory") pod "7928fef2-025b-4bab-ab81-839abc70715b" (UID: "7928fef2-025b-4bab-ab81-839abc70715b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.627596 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.627643 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.627658 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwzcd\" (UniqueName: \"kubernetes.io/projected/7928fef2-025b-4bab-ab81-839abc70715b-kube-api-access-hwzcd\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.627671 4751 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7928fef2-025b-4bab-ab81-839abc70715b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.979707 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" event={"ID":"7928fef2-025b-4bab-ab81-839abc70715b","Type":"ContainerDied","Data":"9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2"} Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.980083 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b777d01bbab32cc63621296f2741af96515633ede1c200861d2b670fa59faf2" Oct 02 11:21:09 crc kubenswrapper[4751]: I1002 11:21:09.979805 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.067992 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6"] Oct 02 11:21:10 crc kubenswrapper[4751]: E1002 11:21:10.068583 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7928fef2-025b-4bab-ab81-839abc70715b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.068688 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7928fef2-025b-4bab-ab81-839abc70715b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.068929 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7928fef2-025b-4bab-ab81-839abc70715b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.069576 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.077841 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.078043 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.077867 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.078002 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.082285 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6"] Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.137333 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9sfl\" (UniqueName: \"kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.137427 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.137737 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.239110 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.239249 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9sfl\" (UniqueName: \"kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.239317 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.243822 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.244754 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.256936 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9sfl\" (UniqueName: \"kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mqgg6\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.388884 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.887663 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6"] Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.894972 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:21:10 crc kubenswrapper[4751]: I1002 11:21:10.990760 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" event={"ID":"bb40236f-9313-40db-9e49-ffe654fdd92e","Type":"ContainerStarted","Data":"9112479efa78c37f40f2e8274b49d6db624dc33e4e7943bfc46f9f868fe094da"} Oct 02 11:21:12 crc kubenswrapper[4751]: I1002 11:21:12.002895 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" event={"ID":"bb40236f-9313-40db-9e49-ffe654fdd92e","Type":"ContainerStarted","Data":"ffe701ca7023215a6f9861ace12327eb7f55713822e8f9364a8a25aa1f5d710a"} Oct 02 11:21:12 crc kubenswrapper[4751]: I1002 11:21:12.030739 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" podStartSLOduration=1.761177375 podStartE2EDuration="2.030717269s" podCreationTimestamp="2025-10-02 11:21:10 +0000 UTC" firstStartedPulling="2025-10-02 11:21:10.894687394 +0000 UTC m=+1752.948913844" lastFinishedPulling="2025-10-02 11:21:11.164227288 +0000 UTC m=+1753.218453738" observedRunningTime="2025-10-02 11:21:12.025098668 +0000 UTC m=+1754.079325148" watchObservedRunningTime="2025-10-02 11:21:12.030717269 +0000 UTC m=+1754.084943729" Oct 02 11:21:14 crc kubenswrapper[4751]: I1002 11:21:14.550444 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:21:14 crc kubenswrapper[4751]: E1002 11:21:14.550917 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:21:15 crc kubenswrapper[4751]: I1002 11:21:15.032595 4751 generic.go:334] "Generic (PLEG): container finished" podID="bb40236f-9313-40db-9e49-ffe654fdd92e" containerID="ffe701ca7023215a6f9861ace12327eb7f55713822e8f9364a8a25aa1f5d710a" exitCode=0 Oct 02 11:21:15 crc kubenswrapper[4751]: I1002 11:21:15.032637 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" event={"ID":"bb40236f-9313-40db-9e49-ffe654fdd92e","Type":"ContainerDied","Data":"ffe701ca7023215a6f9861ace12327eb7f55713822e8f9364a8a25aa1f5d710a"} Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.442318 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.558824 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9sfl\" (UniqueName: \"kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl\") pod \"bb40236f-9313-40db-9e49-ffe654fdd92e\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.559096 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory\") pod \"bb40236f-9313-40db-9e49-ffe654fdd92e\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.559281 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key\") pod \"bb40236f-9313-40db-9e49-ffe654fdd92e\" (UID: \"bb40236f-9313-40db-9e49-ffe654fdd92e\") " Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.564452 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl" (OuterVolumeSpecName: "kube-api-access-c9sfl") pod "bb40236f-9313-40db-9e49-ffe654fdd92e" (UID: "bb40236f-9313-40db-9e49-ffe654fdd92e"). InnerVolumeSpecName "kube-api-access-c9sfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.602348 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory" (OuterVolumeSpecName: "inventory") pod "bb40236f-9313-40db-9e49-ffe654fdd92e" (UID: "bb40236f-9313-40db-9e49-ffe654fdd92e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.625402 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bb40236f-9313-40db-9e49-ffe654fdd92e" (UID: "bb40236f-9313-40db-9e49-ffe654fdd92e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.663975 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.664010 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb40236f-9313-40db-9e49-ffe654fdd92e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:16 crc kubenswrapper[4751]: I1002 11:21:16.664022 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9sfl\" (UniqueName: \"kubernetes.io/projected/bb40236f-9313-40db-9e49-ffe654fdd92e-kube-api-access-c9sfl\") on node \"crc\" DevicePath \"\"" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.054031 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" event={"ID":"bb40236f-9313-40db-9e49-ffe654fdd92e","Type":"ContainerDied","Data":"9112479efa78c37f40f2e8274b49d6db624dc33e4e7943bfc46f9f868fe094da"} Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.054079 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9112479efa78c37f40f2e8274b49d6db624dc33e4e7943bfc46f9f868fe094da" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.054162 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mqgg6" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.134953 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9"] Oct 02 11:21:17 crc kubenswrapper[4751]: E1002 11:21:17.135367 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb40236f-9313-40db-9e49-ffe654fdd92e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.135385 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb40236f-9313-40db-9e49-ffe654fdd92e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.135573 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb40236f-9313-40db-9e49-ffe654fdd92e" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.136228 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.139952 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.140184 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.142818 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.142845 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.148852 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9"] Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.275002 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.275111 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.275134 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj9mq\" (UniqueName: \"kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.275281 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.377292 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.377353 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.377406 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.377426 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj9mq\" (UniqueName: \"kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.381572 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.381795 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.382134 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.397664 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj9mq\" (UniqueName: \"kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.453299 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:21:17 crc kubenswrapper[4751]: I1002 11:21:17.986859 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9"] Oct 02 11:21:18 crc kubenswrapper[4751]: I1002 11:21:18.063408 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" event={"ID":"fe7b2e26-7acf-4950-b080-7f5c5e866491","Type":"ContainerStarted","Data":"7f3f348736153e8d29055a76e5a0b79ddc054e53cd9cb806973afc32ee7ce058"} Oct 02 11:21:19 crc kubenswrapper[4751]: I1002 11:21:19.075683 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" event={"ID":"fe7b2e26-7acf-4950-b080-7f5c5e866491","Type":"ContainerStarted","Data":"d6ddbafcd6f6d67bce2b6415cce847525b01a3d41b3f1c95af1a8f8807bea656"} Oct 02 11:21:19 crc kubenswrapper[4751]: I1002 11:21:19.095600 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" podStartSLOduration=1.920579535 podStartE2EDuration="2.095577707s" podCreationTimestamp="2025-10-02 11:21:17 +0000 UTC" firstStartedPulling="2025-10-02 11:21:17.98909817 +0000 UTC m=+1760.043324630" lastFinishedPulling="2025-10-02 11:21:18.164096342 +0000 UTC m=+1760.218322802" observedRunningTime="2025-10-02 11:21:19.089603856 +0000 UTC m=+1761.143830326" watchObservedRunningTime="2025-10-02 11:21:19.095577707 +0000 UTC m=+1761.149804167" Oct 02 11:21:27 crc kubenswrapper[4751]: I1002 11:21:27.551263 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:21:27 crc kubenswrapper[4751]: E1002 11:21:27.552113 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:21:44 crc kubenswrapper[4751]: I1002 11:21:38.550434 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:21:44 crc kubenswrapper[4751]: E1002 11:21:38.551431 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:21:51 crc kubenswrapper[4751]: I1002 11:21:51.550048 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:21:51 crc kubenswrapper[4751]: E1002 11:21:51.551126 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:22:05 crc kubenswrapper[4751]: I1002 11:22:05.550048 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:22:05 crc kubenswrapper[4751]: E1002 11:22:05.550822 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:22:08 crc kubenswrapper[4751]: I1002 11:22:08.621712 4751 scope.go:117] "RemoveContainer" containerID="4da612369578da04394d2577ad56d3f64936b0b1e67eca23f4cf4d3e2eed7d51" Oct 02 11:22:08 crc kubenswrapper[4751]: I1002 11:22:08.644108 4751 scope.go:117] "RemoveContainer" containerID="4e937be64e941642a19a4590d6a8723cfc3aa2b7aefe489fa23a72b8dd421c05" Oct 02 11:22:20 crc kubenswrapper[4751]: I1002 11:22:20.550398 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:22:20 crc kubenswrapper[4751]: E1002 11:22:20.551273 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:22:35 crc kubenswrapper[4751]: I1002 11:22:35.550597 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:22:35 crc kubenswrapper[4751]: E1002 11:22:35.551200 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:22:46 crc kubenswrapper[4751]: I1002 11:22:46.550117 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:22:46 crc kubenswrapper[4751]: E1002 11:22:46.550816 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:22:59 crc kubenswrapper[4751]: I1002 11:22:59.555521 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:22:59 crc kubenswrapper[4751]: E1002 11:22:59.556230 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:23:08 crc kubenswrapper[4751]: I1002 11:23:08.697942 4751 scope.go:117] "RemoveContainer" containerID="27df7f2fe900c6633ea2b661390517c862083e6b4723546a632198ee3c9e029b" Oct 02 11:23:08 crc kubenswrapper[4751]: I1002 11:23:08.723734 4751 scope.go:117] "RemoveContainer" containerID="735e5ebf324f2df317859ff894db4da9cc4f2cb1ec4fe2641130c89942d11f15" Oct 02 11:23:10 crc kubenswrapper[4751]: I1002 11:23:10.550304 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:23:10 crc kubenswrapper[4751]: E1002 11:23:10.550609 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:23:24 crc kubenswrapper[4751]: I1002 11:23:24.550800 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:23:24 crc kubenswrapper[4751]: E1002 11:23:24.551675 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:23:36 crc kubenswrapper[4751]: I1002 11:23:36.550524 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:23:36 crc kubenswrapper[4751]: E1002 11:23:36.551883 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:23:47 crc kubenswrapper[4751]: I1002 11:23:47.549815 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:23:47 crc kubenswrapper[4751]: E1002 11:23:47.551947 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:24:01 crc kubenswrapper[4751]: I1002 11:24:01.550269 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:24:02 crc kubenswrapper[4751]: I1002 11:24:02.621291 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34"} Oct 02 11:24:08 crc kubenswrapper[4751]: I1002 11:24:08.783501 4751 scope.go:117] "RemoveContainer" containerID="e4995510d541c879a0992ef0e05126af02aead27d8708e9e0daba361c76661b4" Oct 02 11:24:08 crc kubenswrapper[4751]: I1002 11:24:08.818490 4751 scope.go:117] "RemoveContainer" containerID="c67cd3ef1c4fe61ca5288a671379d49fa8971ba5c1aa63aebf51fa029ec17f8a" Oct 02 11:24:08 crc kubenswrapper[4751]: I1002 11:24:08.838239 4751 scope.go:117] "RemoveContainer" containerID="9a086cc41a2751992b4d40879db5a227aecde8b5e959356e7f30d81041832ceb" Oct 02 11:24:08 crc kubenswrapper[4751]: I1002 11:24:08.856822 4751 scope.go:117] "RemoveContainer" containerID="ce301b246e99c191f9083946459cef1b1fbe155015e8dbafca08ea162a31e232" Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.053933 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-fg7cq"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.064010 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qzv47"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.073903 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-ffs5s"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.084530 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-ffs5s"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.091479 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-fg7cq"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.098156 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qzv47"] Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.561519 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39a3d249-497a-4ea8-9752-6f21d35f392d" path="/var/lib/kubelet/pods/39a3d249-497a-4ea8-9752-6f21d35f392d/volumes" Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.562071 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6514a959-a0fd-4226-8676-00c411e7e1fb" path="/var/lib/kubelet/pods/6514a959-a0fd-4226-8676-00c411e7e1fb/volumes" Oct 02 11:24:21 crc kubenswrapper[4751]: I1002 11:24:21.562600 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc55b82d-1e7c-4036-b372-e881a31dcbe5" path="/var/lib/kubelet/pods/bc55b82d-1e7c-4036-b372-e881a31dcbe5/volumes" Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.026711 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ll6rj"] Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.035733 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-l56lv"] Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.046567 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-l56lv"] Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.053206 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ll6rj"] Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.058965 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-pz7qh"] Oct 02 11:24:22 crc kubenswrapper[4751]: I1002 11:24:22.064594 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-pz7qh"] Oct 02 11:24:23 crc kubenswrapper[4751]: I1002 11:24:23.562010 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16d45636-f990-47b9-923f-f1464ea8c046" path="/var/lib/kubelet/pods/16d45636-f990-47b9-923f-f1464ea8c046/volumes" Oct 02 11:24:23 crc kubenswrapper[4751]: I1002 11:24:23.562748 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45165ce-0577-4093-b5b2-268035e7540f" path="/var/lib/kubelet/pods/d45165ce-0577-4093-b5b2-268035e7540f/volumes" Oct 02 11:24:23 crc kubenswrapper[4751]: I1002 11:24:23.563318 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9944c9c-cead-44e6-b589-788b4ef04cc7" path="/var/lib/kubelet/pods/d9944c9c-cead-44e6-b589-788b4ef04cc7/volumes" Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.039853 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ea01-account-create-g628k"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.050620 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-71a7-account-create-nr8bd"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.059253 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3e12-account-create-jvktf"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.067206 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-71a7-account-create-nr8bd"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.076546 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3e12-account-create-jvktf"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.085937 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ea01-account-create-g628k"] Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.561986 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="478f1afd-b217-42d6-b69e-3f66c7a893bf" path="/var/lib/kubelet/pods/478f1afd-b217-42d6-b69e-3f66c7a893bf/volumes" Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.562836 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87713887-97a6-40cc-8c3b-b92b64a26112" path="/var/lib/kubelet/pods/87713887-97a6-40cc-8c3b-b92b64a26112/volumes" Oct 02 11:24:31 crc kubenswrapper[4751]: I1002 11:24:31.563351 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9abde181-ba76-4a0e-9346-af67fcd274f2" path="/var/lib/kubelet/pods/9abde181-ba76-4a0e-9346-af67fcd274f2/volumes" Oct 02 11:24:33 crc kubenswrapper[4751]: I1002 11:24:33.025201 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-df68-account-create-wjhj7"] Oct 02 11:24:33 crc kubenswrapper[4751]: I1002 11:24:33.033012 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-df68-account-create-wjhj7"] Oct 02 11:24:33 crc kubenswrapper[4751]: I1002 11:24:33.569898 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea9cc43-9a45-4a96-9efb-5fb84a515625" path="/var/lib/kubelet/pods/bea9cc43-9a45-4a96-9efb-5fb84a515625/volumes" Oct 02 11:24:34 crc kubenswrapper[4751]: I1002 11:24:34.026274 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-dccd-account-create-pj5cq"] Oct 02 11:24:34 crc kubenswrapper[4751]: I1002 11:24:34.033701 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-afa2-account-create-x7p58"] Oct 02 11:24:34 crc kubenswrapper[4751]: I1002 11:24:34.045823 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-afa2-account-create-x7p58"] Oct 02 11:24:34 crc kubenswrapper[4751]: I1002 11:24:34.056347 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-dccd-account-create-pj5cq"] Oct 02 11:24:35 crc kubenswrapper[4751]: I1002 11:24:35.559898 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af8ba996-86ce-4ded-9615-77d93533e33f" path="/var/lib/kubelet/pods/af8ba996-86ce-4ded-9615-77d93533e33f/volumes" Oct 02 11:24:35 crc kubenswrapper[4751]: I1002 11:24:35.560784 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beeb2d4a-d95e-42d7-98bc-13c1b23ea096" path="/var/lib/kubelet/pods/beeb2d4a-d95e-42d7-98bc-13c1b23ea096/volumes" Oct 02 11:24:53 crc kubenswrapper[4751]: I1002 11:24:53.037993 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-v4zhn"] Oct 02 11:24:53 crc kubenswrapper[4751]: I1002 11:24:53.053112 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-v4zhn"] Oct 02 11:24:53 crc kubenswrapper[4751]: I1002 11:24:53.561768 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6df6e4-d37d-4862-9f23-4932c55784ec" path="/var/lib/kubelet/pods/ff6df6e4-d37d-4862-9f23-4932c55784ec/volumes" Oct 02 11:25:08 crc kubenswrapper[4751]: I1002 11:25:08.913346 4751 scope.go:117] "RemoveContainer" containerID="e96ef46ae2fdba975bd6990dbc6007d8d60433914e44c082acaa0920958de230" Oct 02 11:25:08 crc kubenswrapper[4751]: I1002 11:25:08.943572 4751 scope.go:117] "RemoveContainer" containerID="887ad8f641f8fe949619cff876c524efe0bb8ecac2d1dcfb83762f34cb0e2605" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.032482 4751 scope.go:117] "RemoveContainer" containerID="60bf10fbda72722ef2d9f5aff822f3ba8be4375b9295fef422e24fb517c2ff39" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.063464 4751 scope.go:117] "RemoveContainer" containerID="472b02f690b09f41d9dcb9a58d7c2cfc58909fdf059e91c3fdf9bf36db484b2f" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.155372 4751 scope.go:117] "RemoveContainer" containerID="a0c97fd3714f56ccdd60f3d1fcd149aa9d49120281592d23f787660f554b7565" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.185698 4751 scope.go:117] "RemoveContainer" containerID="5a79a95a25cc83a83077fd6a66d465e5c19b052756122260a117d710e9e80930" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.217405 4751 scope.go:117] "RemoveContainer" containerID="0cc9c5351195fb28145a8f3243beba900678730dd228b5fecb6e920247eeda0e" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.243497 4751 scope.go:117] "RemoveContainer" containerID="720bf7cc015540cca238cc11156072acaec8b5c7cf306a86feeb69781eb45d13" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.295754 4751 scope.go:117] "RemoveContainer" containerID="3e37fb9ae9ba706d2e7582dbff72766d97ba15cb4e3c95979d821db91651455f" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.313104 4751 scope.go:117] "RemoveContainer" containerID="18562bd55ef9ea77c31cbc410d5cb13f30ee95cb01dcb15a076a0f6f573fb1ea" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.353652 4751 scope.go:117] "RemoveContainer" containerID="504b9d94d0a5bc36891f30d0cf058fff09bbea782297045d58a32769c1b9dd2f" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.375882 4751 scope.go:117] "RemoveContainer" containerID="68db4c24bc1909258a77a59fbebb97eb8008cd8e86f4a31e31bbc33b18031ec9" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.395014 4751 scope.go:117] "RemoveContainer" containerID="946b76220a37bd82f78c15f93a3a24ad4369dce8671dc3ebf96c66fecd7ea82e" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.418003 4751 scope.go:117] "RemoveContainer" containerID="d137206074fbe87568b54ac62fe82a923f225ef4f3e0ca1c32d3ff70993dfad0" Oct 02 11:25:09 crc kubenswrapper[4751]: I1002 11:25:09.452763 4751 scope.go:117] "RemoveContainer" containerID="ee670e75c1498a1a22998d51629e39aca6decdcb1078ff1ad1904f08bf90eb5e" Oct 02 11:25:22 crc kubenswrapper[4751]: I1002 11:25:22.334843 4751 generic.go:334] "Generic (PLEG): container finished" podID="fe7b2e26-7acf-4950-b080-7f5c5e866491" containerID="d6ddbafcd6f6d67bce2b6415cce847525b01a3d41b3f1c95af1a8f8807bea656" exitCode=0 Oct 02 11:25:22 crc kubenswrapper[4751]: I1002 11:25:22.334947 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" event={"ID":"fe7b2e26-7acf-4950-b080-7f5c5e866491","Type":"ContainerDied","Data":"d6ddbafcd6f6d67bce2b6415cce847525b01a3d41b3f1c95af1a8f8807bea656"} Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.756903 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.796617 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory\") pod \"fe7b2e26-7acf-4950-b080-7f5c5e866491\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.796752 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj9mq\" (UniqueName: \"kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq\") pod \"fe7b2e26-7acf-4950-b080-7f5c5e866491\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.796799 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle\") pod \"fe7b2e26-7acf-4950-b080-7f5c5e866491\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.796896 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key\") pod \"fe7b2e26-7acf-4950-b080-7f5c5e866491\" (UID: \"fe7b2e26-7acf-4950-b080-7f5c5e866491\") " Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.802948 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq" (OuterVolumeSpecName: "kube-api-access-bj9mq") pod "fe7b2e26-7acf-4950-b080-7f5c5e866491" (UID: "fe7b2e26-7acf-4950-b080-7f5c5e866491"). InnerVolumeSpecName "kube-api-access-bj9mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.803162 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fe7b2e26-7acf-4950-b080-7f5c5e866491" (UID: "fe7b2e26-7acf-4950-b080-7f5c5e866491"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.829214 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fe7b2e26-7acf-4950-b080-7f5c5e866491" (UID: "fe7b2e26-7acf-4950-b080-7f5c5e866491"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.831970 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory" (OuterVolumeSpecName: "inventory") pod "fe7b2e26-7acf-4950-b080-7f5c5e866491" (UID: "fe7b2e26-7acf-4950-b080-7f5c5e866491"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.899511 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.899575 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj9mq\" (UniqueName: \"kubernetes.io/projected/fe7b2e26-7acf-4950-b080-7f5c5e866491-kube-api-access-bj9mq\") on node \"crc\" DevicePath \"\"" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.899598 4751 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:25:23 crc kubenswrapper[4751]: I1002 11:25:23.899610 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe7b2e26-7acf-4950-b080-7f5c5e866491-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.353884 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" event={"ID":"fe7b2e26-7acf-4950-b080-7f5c5e866491","Type":"ContainerDied","Data":"7f3f348736153e8d29055a76e5a0b79ddc054e53cd9cb806973afc32ee7ce058"} Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.353918 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f3f348736153e8d29055a76e5a0b79ddc054e53cd9cb806973afc32ee7ce058" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.353934 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.465042 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l"] Oct 02 11:25:24 crc kubenswrapper[4751]: E1002 11:25:24.465563 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b2e26-7acf-4950-b080-7f5c5e866491" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.465588 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b2e26-7acf-4950-b080-7f5c5e866491" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.465805 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b2e26-7acf-4950-b080-7f5c5e866491" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.466504 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.468931 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.469186 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.469345 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.469726 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.478902 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l"] Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.609801 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4xbn\" (UniqueName: \"kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.610010 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.610073 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.711704 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4xbn\" (UniqueName: \"kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.712119 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.712195 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.716347 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.718648 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.730549 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4xbn\" (UniqueName: \"kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:24 crc kubenswrapper[4751]: I1002 11:25:24.788871 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:25:25 crc kubenswrapper[4751]: I1002 11:25:25.280293 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l"] Oct 02 11:25:25 crc kubenswrapper[4751]: I1002 11:25:25.363485 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" event={"ID":"ffb43401-318b-43e5-b381-01d9bbf445cc","Type":"ContainerStarted","Data":"b2ae47cea70394fb7907d64f8e1bdd05e5b7e37ae64169c130edea7ed9512616"} Oct 02 11:25:26 crc kubenswrapper[4751]: I1002 11:25:26.374546 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" event={"ID":"ffb43401-318b-43e5-b381-01d9bbf445cc","Type":"ContainerStarted","Data":"683e90692e8c35c71755ee4c792a91d28ea733be86c4bb0f856f8ce708a2b1a4"} Oct 02 11:25:26 crc kubenswrapper[4751]: I1002 11:25:26.400229 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" podStartSLOduration=2.208100338 podStartE2EDuration="2.400209136s" podCreationTimestamp="2025-10-02 11:25:24 +0000 UTC" firstStartedPulling="2025-10-02 11:25:25.287986705 +0000 UTC m=+2007.342213155" lastFinishedPulling="2025-10-02 11:25:25.480095503 +0000 UTC m=+2007.534321953" observedRunningTime="2025-10-02 11:25:26.391608826 +0000 UTC m=+2008.445835286" watchObservedRunningTime="2025-10-02 11:25:26.400209136 +0000 UTC m=+2008.454435596" Oct 02 11:25:52 crc kubenswrapper[4751]: I1002 11:25:52.038478 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-44kgc"] Oct 02 11:25:52 crc kubenswrapper[4751]: I1002 11:25:52.047933 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-44kgc"] Oct 02 11:25:53 crc kubenswrapper[4751]: I1002 11:25:53.035512 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fbknj"] Oct 02 11:25:53 crc kubenswrapper[4751]: I1002 11:25:53.048920 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fbknj"] Oct 02 11:25:53 crc kubenswrapper[4751]: I1002 11:25:53.562615 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d2a342d-289c-41e9-94f7-6ddbc10a6e1c" path="/var/lib/kubelet/pods/4d2a342d-289c-41e9-94f7-6ddbc10a6e1c/volumes" Oct 02 11:25:53 crc kubenswrapper[4751]: I1002 11:25:53.563598 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f251a17b-99b0-4656-919b-1d03af5331a4" path="/var/lib/kubelet/pods/f251a17b-99b0-4656-919b-1d03af5331a4/volumes" Oct 02 11:25:55 crc kubenswrapper[4751]: I1002 11:25:55.033944 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-jmv2n"] Oct 02 11:25:55 crc kubenswrapper[4751]: I1002 11:25:55.044536 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-jmv2n"] Oct 02 11:25:55 crc kubenswrapper[4751]: I1002 11:25:55.563800 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a15cb7-4b3b-4b25-8b11-1b8f14b06479" path="/var/lib/kubelet/pods/d0a15cb7-4b3b-4b25-8b11-1b8f14b06479/volumes" Oct 02 11:26:01 crc kubenswrapper[4751]: I1002 11:26:01.507194 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:26:01 crc kubenswrapper[4751]: I1002 11:26:01.508162 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:26:09 crc kubenswrapper[4751]: I1002 11:26:09.780524 4751 scope.go:117] "RemoveContainer" containerID="119e0a5ecc875d6dce924f32efe9df3e575a54fe719c0861879bebe767dff91f" Oct 02 11:26:09 crc kubenswrapper[4751]: I1002 11:26:09.810849 4751 scope.go:117] "RemoveContainer" containerID="50836859af1ede94a8266a964e92ff2894489e3be518f679a8fe934fb90b06e4" Oct 02 11:26:09 crc kubenswrapper[4751]: I1002 11:26:09.876805 4751 scope.go:117] "RemoveContainer" containerID="52cc443f91d71eb36ec985ec6a87d5dd4b78c443e81219875e00519e8f937532" Oct 02 11:26:22 crc kubenswrapper[4751]: I1002 11:26:22.042443 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-247rg"] Oct 02 11:26:22 crc kubenswrapper[4751]: I1002 11:26:22.049067 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-247rg"] Oct 02 11:26:23 crc kubenswrapper[4751]: I1002 11:26:23.566777 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9bb5b1a-3008-4ca4-8c88-dac1b5b22471" path="/var/lib/kubelet/pods/c9bb5b1a-3008-4ca4-8c88-dac1b5b22471/volumes" Oct 02 11:26:24 crc kubenswrapper[4751]: I1002 11:26:24.028615 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-m2fxt"] Oct 02 11:26:24 crc kubenswrapper[4751]: I1002 11:26:24.036965 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-m2fxt"] Oct 02 11:26:25 crc kubenswrapper[4751]: I1002 11:26:25.567975 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6080e003-0583-428a-b263-3badfde0fd37" path="/var/lib/kubelet/pods/6080e003-0583-428a-b263-3badfde0fd37/volumes" Oct 02 11:26:30 crc kubenswrapper[4751]: I1002 11:26:30.037355 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-28csr"] Oct 02 11:26:30 crc kubenswrapper[4751]: I1002 11:26:30.044757 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-28csr"] Oct 02 11:26:31 crc kubenswrapper[4751]: I1002 11:26:31.507342 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:26:31 crc kubenswrapper[4751]: I1002 11:26:31.507715 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:26:31 crc kubenswrapper[4751]: I1002 11:26:31.569433 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faae59de-da62-46ae-b58e-acb43ada5fec" path="/var/lib/kubelet/pods/faae59de-da62-46ae-b58e-acb43ada5fec/volumes" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.038731 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-g7q7n"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.048262 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-m8dqc"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.061761 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-hfxcj"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.070877 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-m8dqc"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.079775 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-g7q7n"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.087366 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-hfxcj"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.608358 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.612583 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.622451 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.773682 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwhtm\" (UniqueName: \"kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.773934 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.774058 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.875742 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwhtm\" (UniqueName: \"kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.875843 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.875876 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.876576 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.876844 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.903472 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwhtm\" (UniqueName: \"kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm\") pod \"community-operators-zjkgh\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:46 crc kubenswrapper[4751]: I1002 11:26:46.939054 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:47 crc kubenswrapper[4751]: I1002 11:26:47.437766 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:26:47 crc kubenswrapper[4751]: I1002 11:26:47.568009 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af13930-d981-4282-831e-aa071b1e1e85" path="/var/lib/kubelet/pods/0af13930-d981-4282-831e-aa071b1e1e85/volumes" Oct 02 11:26:47 crc kubenswrapper[4751]: I1002 11:26:47.569049 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c468c9e-b55c-4298-95fd-b382f454f5e6" path="/var/lib/kubelet/pods/2c468c9e-b55c-4298-95fd-b382f454f5e6/volumes" Oct 02 11:26:47 crc kubenswrapper[4751]: I1002 11:26:47.569975 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca" path="/var/lib/kubelet/pods/d8c5f5c3-f2a5-4b95-9e64-6c3f02db19ca/volumes" Oct 02 11:26:48 crc kubenswrapper[4751]: I1002 11:26:48.126770 4751 generic.go:334] "Generic (PLEG): container finished" podID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerID="8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766" exitCode=0 Oct 02 11:26:48 crc kubenswrapper[4751]: I1002 11:26:48.126856 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerDied","Data":"8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766"} Oct 02 11:26:48 crc kubenswrapper[4751]: I1002 11:26:48.127115 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerStarted","Data":"2c963fb271f16659bd45a15512fd05656a72a5cbbad37711bb39134f439a787d"} Oct 02 11:26:48 crc kubenswrapper[4751]: I1002 11:26:48.129589 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:26:50 crc kubenswrapper[4751]: I1002 11:26:50.145071 4751 generic.go:334] "Generic (PLEG): container finished" podID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerID="63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df" exitCode=0 Oct 02 11:26:50 crc kubenswrapper[4751]: I1002 11:26:50.145218 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerDied","Data":"63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df"} Oct 02 11:26:51 crc kubenswrapper[4751]: I1002 11:26:51.156387 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerStarted","Data":"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0"} Oct 02 11:26:51 crc kubenswrapper[4751]: I1002 11:26:51.179487 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zjkgh" podStartSLOduration=2.6919989810000002 podStartE2EDuration="5.179469437s" podCreationTimestamp="2025-10-02 11:26:46 +0000 UTC" firstStartedPulling="2025-10-02 11:26:48.129127196 +0000 UTC m=+2090.183353666" lastFinishedPulling="2025-10-02 11:26:50.616597672 +0000 UTC m=+2092.670824122" observedRunningTime="2025-10-02 11:26:51.173780594 +0000 UTC m=+2093.228007044" watchObservedRunningTime="2025-10-02 11:26:51.179469437 +0000 UTC m=+2093.233695887" Oct 02 11:26:56 crc kubenswrapper[4751]: I1002 11:26:56.939611 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:56 crc kubenswrapper[4751]: I1002 11:26:56.940304 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:56 crc kubenswrapper[4751]: I1002 11:26:56.985941 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:57 crc kubenswrapper[4751]: I1002 11:26:57.033147 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-551e-account-create-pw5vt"] Oct 02 11:26:57 crc kubenswrapper[4751]: I1002 11:26:57.044926 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-551e-account-create-pw5vt"] Oct 02 11:26:57 crc kubenswrapper[4751]: I1002 11:26:57.256197 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:57 crc kubenswrapper[4751]: I1002 11:26:57.304951 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:26:57 crc kubenswrapper[4751]: I1002 11:26:57.567277 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2319b758-f6a7-4d3f-b1c8-25747580a255" path="/var/lib/kubelet/pods/2319b758-f6a7-4d3f-b1c8-25747580a255/volumes" Oct 02 11:26:58 crc kubenswrapper[4751]: I1002 11:26:58.028948 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e375-account-create-7d49s"] Oct 02 11:26:58 crc kubenswrapper[4751]: I1002 11:26:58.035789 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e375-account-create-7d49s"] Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.035292 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a8c7-account-create-cxz7f"] Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.044973 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a8c7-account-create-cxz7f"] Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.224677 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zjkgh" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="registry-server" containerID="cri-o://3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0" gracePeriod=2 Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.561428 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e8d5277-79db-4340-a8fa-66b661abd20a" path="/var/lib/kubelet/pods/7e8d5277-79db-4340-a8fa-66b661abd20a/volumes" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.562389 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="939a3dba-d15a-4257-987d-c931791cf416" path="/var/lib/kubelet/pods/939a3dba-d15a-4257-987d-c931791cf416/volumes" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.676622 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.823511 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content\") pod \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.823607 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities\") pod \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.823755 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwhtm\" (UniqueName: \"kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm\") pod \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\" (UID: \"6cd8d994-4684-412d-8f4a-36a3c4fd2908\") " Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.824811 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities" (OuterVolumeSpecName: "utilities") pod "6cd8d994-4684-412d-8f4a-36a3c4fd2908" (UID: "6cd8d994-4684-412d-8f4a-36a3c4fd2908"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.828613 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm" (OuterVolumeSpecName: "kube-api-access-zwhtm") pod "6cd8d994-4684-412d-8f4a-36a3c4fd2908" (UID: "6cd8d994-4684-412d-8f4a-36a3c4fd2908"). InnerVolumeSpecName "kube-api-access-zwhtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.873869 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cd8d994-4684-412d-8f4a-36a3c4fd2908" (UID: "6cd8d994-4684-412d-8f4a-36a3c4fd2908"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.926017 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.926067 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd8d994-4684-412d-8f4a-36a3c4fd2908-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:26:59 crc kubenswrapper[4751]: I1002 11:26:59.926086 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwhtm\" (UniqueName: \"kubernetes.io/projected/6cd8d994-4684-412d-8f4a-36a3c4fd2908-kube-api-access-zwhtm\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.238833 4751 generic.go:334] "Generic (PLEG): container finished" podID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerID="3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0" exitCode=0 Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.238901 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerDied","Data":"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0"} Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.238908 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zjkgh" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.238950 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zjkgh" event={"ID":"6cd8d994-4684-412d-8f4a-36a3c4fd2908","Type":"ContainerDied","Data":"2c963fb271f16659bd45a15512fd05656a72a5cbbad37711bb39134f439a787d"} Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.238981 4751 scope.go:117] "RemoveContainer" containerID="3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.261384 4751 scope.go:117] "RemoveContainer" containerID="63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.281322 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.292818 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zjkgh"] Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.320745 4751 scope.go:117] "RemoveContainer" containerID="8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.391060 4751 scope.go:117] "RemoveContainer" containerID="3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0" Oct 02 11:27:00 crc kubenswrapper[4751]: E1002 11:27:00.391573 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0\": container with ID starting with 3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0 not found: ID does not exist" containerID="3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.391684 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0"} err="failed to get container status \"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0\": rpc error: code = NotFound desc = could not find container \"3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0\": container with ID starting with 3b2953b823a1e125e5f15840f4db6407c8fa180e30382611e6982c5afb9057e0 not found: ID does not exist" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.391791 4751 scope.go:117] "RemoveContainer" containerID="63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df" Oct 02 11:27:00 crc kubenswrapper[4751]: E1002 11:27:00.392523 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df\": container with ID starting with 63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df not found: ID does not exist" containerID="63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.392602 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df"} err="failed to get container status \"63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df\": rpc error: code = NotFound desc = could not find container \"63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df\": container with ID starting with 63b7103a0dc6e540c84f08ccbf18985eaeff160c50b7134b0e8fc6132add06df not found: ID does not exist" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.392674 4751 scope.go:117] "RemoveContainer" containerID="8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766" Oct 02 11:27:00 crc kubenswrapper[4751]: E1002 11:27:00.393041 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766\": container with ID starting with 8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766 not found: ID does not exist" containerID="8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766" Oct 02 11:27:00 crc kubenswrapper[4751]: I1002 11:27:00.393090 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766"} err="failed to get container status \"8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766\": rpc error: code = NotFound desc = could not find container \"8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766\": container with ID starting with 8b33ea36aab281805cad80eb3bbefa309ae7bebbf233597c22d4d308f1662766 not found: ID does not exist" Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.507364 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.507731 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.507779 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.508551 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.508619 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34" gracePeriod=600 Oct 02 11:27:01 crc kubenswrapper[4751]: I1002 11:27:01.563329 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" path="/var/lib/kubelet/pods/6cd8d994-4684-412d-8f4a-36a3c4fd2908/volumes" Oct 02 11:27:02 crc kubenswrapper[4751]: I1002 11:27:02.259065 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34" exitCode=0 Oct 02 11:27:02 crc kubenswrapper[4751]: I1002 11:27:02.259104 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34"} Oct 02 11:27:02 crc kubenswrapper[4751]: I1002 11:27:02.259498 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f"} Oct 02 11:27:02 crc kubenswrapper[4751]: I1002 11:27:02.259522 4751 scope.go:117] "RemoveContainer" containerID="80153951cecabd988c703df4b14fd7c811e229bcbf8ac53629dc797577156168" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.008673 4751 scope.go:117] "RemoveContainer" containerID="6e4f9dc08829d521d9dca22ef606609231cc5b63de5bc0a108715570307c7fa0" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.037962 4751 scope.go:117] "RemoveContainer" containerID="c18d58b46df5495cfd6327a660967ab882759dc5023dd747eef88056e9093a89" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.115749 4751 scope.go:117] "RemoveContainer" containerID="8703551d47e0a33f9113189ab52914c5416d36420ba174c5d2cada59c8f8dea1" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.166270 4751 scope.go:117] "RemoveContainer" containerID="18f100a45c6dd9cc4f685ce65e2537cb77b6f5f955580996c7c1d1f3fd986a56" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.226868 4751 scope.go:117] "RemoveContainer" containerID="a728fae2f1a380c15e5618cdda55c64f9b462f8bb159e26d29d607b390c303a3" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.292142 4751 scope.go:117] "RemoveContainer" containerID="aacb5b67735a8ffbe6b2aabe3a522c641bf07481cfc35b1f438482ec87a91f8e" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.322634 4751 scope.go:117] "RemoveContainer" containerID="d069b51c33c4aaa6cea9bae8af2f121406780a4b63d32970584f8f09a323e1b1" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.357966 4751 scope.go:117] "RemoveContainer" containerID="2c701b07095f80e46dcb30b2fb96048e33d67f797796f9b03d7ce01770fc0a5b" Oct 02 11:27:10 crc kubenswrapper[4751]: I1002 11:27:10.390301 4751 scope.go:117] "RemoveContainer" containerID="c68f013e7bfc73f4d5edc6719b962c1efc55d54c714c65d771449a895b93e28c" Oct 02 11:27:14 crc kubenswrapper[4751]: I1002 11:27:14.381967 4751 generic.go:334] "Generic (PLEG): container finished" podID="ffb43401-318b-43e5-b381-01d9bbf445cc" containerID="683e90692e8c35c71755ee4c792a91d28ea733be86c4bb0f856f8ce708a2b1a4" exitCode=0 Oct 02 11:27:14 crc kubenswrapper[4751]: I1002 11:27:14.382049 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" event={"ID":"ffb43401-318b-43e5-b381-01d9bbf445cc","Type":"ContainerDied","Data":"683e90692e8c35c71755ee4c792a91d28ea733be86c4bb0f856f8ce708a2b1a4"} Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.801668 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.929412 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory\") pod \"ffb43401-318b-43e5-b381-01d9bbf445cc\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.929671 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4xbn\" (UniqueName: \"kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn\") pod \"ffb43401-318b-43e5-b381-01d9bbf445cc\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.929712 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key\") pod \"ffb43401-318b-43e5-b381-01d9bbf445cc\" (UID: \"ffb43401-318b-43e5-b381-01d9bbf445cc\") " Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.935201 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn" (OuterVolumeSpecName: "kube-api-access-f4xbn") pod "ffb43401-318b-43e5-b381-01d9bbf445cc" (UID: "ffb43401-318b-43e5-b381-01d9bbf445cc"). InnerVolumeSpecName "kube-api-access-f4xbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.966816 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory" (OuterVolumeSpecName: "inventory") pod "ffb43401-318b-43e5-b381-01d9bbf445cc" (UID: "ffb43401-318b-43e5-b381-01d9bbf445cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:27:15 crc kubenswrapper[4751]: I1002 11:27:15.968717 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ffb43401-318b-43e5-b381-01d9bbf445cc" (UID: "ffb43401-318b-43e5-b381-01d9bbf445cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.031652 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4xbn\" (UniqueName: \"kubernetes.io/projected/ffb43401-318b-43e5-b381-01d9bbf445cc-kube-api-access-f4xbn\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.031692 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.031708 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffb43401-318b-43e5-b381-01d9bbf445cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.400052 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" event={"ID":"ffb43401-318b-43e5-b381-01d9bbf445cc","Type":"ContainerDied","Data":"b2ae47cea70394fb7907d64f8e1bdd05e5b7e37ae64169c130edea7ed9512616"} Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.400093 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2ae47cea70394fb7907d64f8e1bdd05e5b7e37ae64169c130edea7ed9512616" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.400132 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.503342 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2"] Oct 02 11:27:16 crc kubenswrapper[4751]: E1002 11:27:16.504038 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb43401-318b-43e5-b381-01d9bbf445cc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504124 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb43401-318b-43e5-b381-01d9bbf445cc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 11:27:16 crc kubenswrapper[4751]: E1002 11:27:16.504221 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="extract-utilities" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504298 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="extract-utilities" Oct 02 11:27:16 crc kubenswrapper[4751]: E1002 11:27:16.504352 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="extract-content" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504436 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="extract-content" Oct 02 11:27:16 crc kubenswrapper[4751]: E1002 11:27:16.504522 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504576 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504886 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd8d994-4684-412d-8f4a-36a3c4fd2908" containerName="registry-server" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.504952 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb43401-318b-43e5-b381-01d9bbf445cc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.505684 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.507854 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.508407 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.510332 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.511366 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.516741 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2"] Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.642828 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.642971 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cbbm\" (UniqueName: \"kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.643061 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.745342 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.745946 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cbbm\" (UniqueName: \"kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.746050 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.752487 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.753657 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.764363 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cbbm\" (UniqueName: \"kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:16 crc kubenswrapper[4751]: I1002 11:27:16.841344 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:27:17 crc kubenswrapper[4751]: I1002 11:27:17.345849 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2"] Oct 02 11:27:17 crc kubenswrapper[4751]: I1002 11:27:17.408552 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" event={"ID":"377f0b22-1de1-49d9-a4d6-f48bdad67770","Type":"ContainerStarted","Data":"0bf04ad4cb3458c068e33f47b48ce3b5e3469beb85b2d5eccc64bacfc8adb2cf"} Oct 02 11:27:18 crc kubenswrapper[4751]: I1002 11:27:18.417477 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" event={"ID":"377f0b22-1de1-49d9-a4d6-f48bdad67770","Type":"ContainerStarted","Data":"bd4d49390baf4bea7a4430c76f8d520a27ad6a5c555238f834f47889116bda8c"} Oct 02 11:27:18 crc kubenswrapper[4751]: I1002 11:27:18.442649 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" podStartSLOduration=1.942015646 podStartE2EDuration="2.442605368s" podCreationTimestamp="2025-10-02 11:27:16 +0000 UTC" firstStartedPulling="2025-10-02 11:27:17.347303806 +0000 UTC m=+2119.401530256" lastFinishedPulling="2025-10-02 11:27:17.847893498 +0000 UTC m=+2119.902119978" observedRunningTime="2025-10-02 11:27:18.442328241 +0000 UTC m=+2120.496554711" watchObservedRunningTime="2025-10-02 11:27:18.442605368 +0000 UTC m=+2120.496831828" Oct 02 11:27:41 crc kubenswrapper[4751]: I1002 11:27:41.041844 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-q7dsc"] Oct 02 11:27:41 crc kubenswrapper[4751]: I1002 11:27:41.050703 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-q7dsc"] Oct 02 11:27:41 crc kubenswrapper[4751]: I1002 11:27:41.561129 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d" path="/var/lib/kubelet/pods/a0a071be-a416-4dc1-91e6-f4d6ed5ddf8d/volumes" Oct 02 11:28:04 crc kubenswrapper[4751]: I1002 11:28:04.041741 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-sq6r8"] Oct 02 11:28:04 crc kubenswrapper[4751]: I1002 11:28:04.048995 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-sq6r8"] Oct 02 11:28:05 crc kubenswrapper[4751]: I1002 11:28:05.565380 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6092dc7b-23dd-4831-a5d7-ad61b6a5d237" path="/var/lib/kubelet/pods/6092dc7b-23dd-4831-a5d7-ad61b6a5d237/volumes" Oct 02 11:28:06 crc kubenswrapper[4751]: I1002 11:28:06.028964 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2nsmc"] Oct 02 11:28:06 crc kubenswrapper[4751]: I1002 11:28:06.036843 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2nsmc"] Oct 02 11:28:07 crc kubenswrapper[4751]: I1002 11:28:07.565999 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e375b8c4-4852-45d4-a6d3-d47afbe12c90" path="/var/lib/kubelet/pods/e375b8c4-4852-45d4-a6d3-d47afbe12c90/volumes" Oct 02 11:28:10 crc kubenswrapper[4751]: I1002 11:28:10.877148 4751 scope.go:117] "RemoveContainer" containerID="dc96115b5cd493c6a2f1188e0cd795d4e1dda9a3e2acd5092634f0880bf29bbc" Oct 02 11:28:10 crc kubenswrapper[4751]: I1002 11:28:10.934605 4751 scope.go:117] "RemoveContainer" containerID="182c3a4c0c5995c735c441f652a5dd7e40f34bb89cd11b97ac14e1d6cac2edde" Oct 02 11:28:10 crc kubenswrapper[4751]: I1002 11:28:10.980304 4751 scope.go:117] "RemoveContainer" containerID="20c9bca24ec08748601b40c0fa0461c7ba783625d330862673abe666392994b9" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.346789 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.349423 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.364739 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.399715 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtxg\" (UniqueName: \"kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.399794 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.399882 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.501562 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtxg\" (UniqueName: \"kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.501696 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.501793 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.502431 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.502455 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.532073 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtxg\" (UniqueName: \"kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg\") pod \"redhat-operators-z2j96\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:32 crc kubenswrapper[4751]: I1002 11:28:32.674774 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:33 crc kubenswrapper[4751]: I1002 11:28:33.151036 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:34 crc kubenswrapper[4751]: I1002 11:28:34.090982 4751 generic.go:334] "Generic (PLEG): container finished" podID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerID="142977f5072151cd599716360061c74b96a5a3f377869230e501df8882e47ec7" exitCode=0 Oct 02 11:28:34 crc kubenswrapper[4751]: I1002 11:28:34.091035 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerDied","Data":"142977f5072151cd599716360061c74b96a5a3f377869230e501df8882e47ec7"} Oct 02 11:28:34 crc kubenswrapper[4751]: I1002 11:28:34.091354 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerStarted","Data":"b2d3404c58d9974edd2fe32b906496c9a7884201d2d5dcab8829df6abd78a3a7"} Oct 02 11:28:35 crc kubenswrapper[4751]: I1002 11:28:35.106998 4751 generic.go:334] "Generic (PLEG): container finished" podID="377f0b22-1de1-49d9-a4d6-f48bdad67770" containerID="bd4d49390baf4bea7a4430c76f8d520a27ad6a5c555238f834f47889116bda8c" exitCode=0 Oct 02 11:28:35 crc kubenswrapper[4751]: I1002 11:28:35.107057 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" event={"ID":"377f0b22-1de1-49d9-a4d6-f48bdad67770","Type":"ContainerDied","Data":"bd4d49390baf4bea7a4430c76f8d520a27ad6a5c555238f834f47889116bda8c"} Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.116660 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerStarted","Data":"86e2f9486e561822c75867e8ac75be15fa21cb27b9103f872c5ad116d7633e5b"} Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.606015 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.678733 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cbbm\" (UniqueName: \"kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm\") pod \"377f0b22-1de1-49d9-a4d6-f48bdad67770\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.678878 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key\") pod \"377f0b22-1de1-49d9-a4d6-f48bdad67770\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.678902 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory\") pod \"377f0b22-1de1-49d9-a4d6-f48bdad67770\" (UID: \"377f0b22-1de1-49d9-a4d6-f48bdad67770\") " Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.684221 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm" (OuterVolumeSpecName: "kube-api-access-8cbbm") pod "377f0b22-1de1-49d9-a4d6-f48bdad67770" (UID: "377f0b22-1de1-49d9-a4d6-f48bdad67770"). InnerVolumeSpecName "kube-api-access-8cbbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.704067 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory" (OuterVolumeSpecName: "inventory") pod "377f0b22-1de1-49d9-a4d6-f48bdad67770" (UID: "377f0b22-1de1-49d9-a4d6-f48bdad67770"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.708393 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "377f0b22-1de1-49d9-a4d6-f48bdad67770" (UID: "377f0b22-1de1-49d9-a4d6-f48bdad67770"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.782481 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cbbm\" (UniqueName: \"kubernetes.io/projected/377f0b22-1de1-49d9-a4d6-f48bdad67770-kube-api-access-8cbbm\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.782545 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:36 crc kubenswrapper[4751]: I1002 11:28:36.782558 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/377f0b22-1de1-49d9-a4d6-f48bdad67770-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.128516 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" event={"ID":"377f0b22-1de1-49d9-a4d6-f48bdad67770","Type":"ContainerDied","Data":"0bf04ad4cb3458c068e33f47b48ce3b5e3469beb85b2d5eccc64bacfc8adb2cf"} Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.128592 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf04ad4cb3458c068e33f47b48ce3b5e3469beb85b2d5eccc64bacfc8adb2cf" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.129274 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.131822 4751 generic.go:334] "Generic (PLEG): container finished" podID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerID="86e2f9486e561822c75867e8ac75be15fa21cb27b9103f872c5ad116d7633e5b" exitCode=0 Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.131888 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerDied","Data":"86e2f9486e561822c75867e8ac75be15fa21cb27b9103f872c5ad116d7633e5b"} Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.217393 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f"] Oct 02 11:28:37 crc kubenswrapper[4751]: E1002 11:28:37.217785 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377f0b22-1de1-49d9-a4d6-f48bdad67770" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.217802 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="377f0b22-1de1-49d9-a4d6-f48bdad67770" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.218008 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="377f0b22-1de1-49d9-a4d6-f48bdad67770" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.218708 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.220660 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.221149 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.221215 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.221834 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.237988 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f"] Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.292571 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.292619 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.292764 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.394319 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.394440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.394465 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.400768 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.400772 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.422062 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:37 crc kubenswrapper[4751]: I1002 11:28:37.538268 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:38 crc kubenswrapper[4751]: I1002 11:28:38.081241 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f"] Oct 02 11:28:38 crc kubenswrapper[4751]: I1002 11:28:38.141286 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" event={"ID":"74695503-3525-42a8-ad11-0e6644eebcee","Type":"ContainerStarted","Data":"48fbccdd79da499b5fcbb50a95636be5c71b0933a6ee8258b9af897041bebb81"} Oct 02 11:28:41 crc kubenswrapper[4751]: I1002 11:28:41.169913 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" event={"ID":"74695503-3525-42a8-ad11-0e6644eebcee","Type":"ContainerStarted","Data":"0c0e3f264a78c9516723637f06a4f974974012d6804d5b8d031fc1efa7b0d0ae"} Oct 02 11:28:41 crc kubenswrapper[4751]: I1002 11:28:41.173639 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerStarted","Data":"e9caf54f7a4791b32a6cb5e4f3136c74c254fa97baab118606806d9b85016594"} Oct 02 11:28:41 crc kubenswrapper[4751]: I1002 11:28:41.199280 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" podStartSLOduration=1.495674945 podStartE2EDuration="4.199251366s" podCreationTimestamp="2025-10-02 11:28:37 +0000 UTC" firstStartedPulling="2025-10-02 11:28:38.090864522 +0000 UTC m=+2200.145090962" lastFinishedPulling="2025-10-02 11:28:40.794440933 +0000 UTC m=+2202.848667383" observedRunningTime="2025-10-02 11:28:41.187904161 +0000 UTC m=+2203.242130651" watchObservedRunningTime="2025-10-02 11:28:41.199251366 +0000 UTC m=+2203.253477846" Oct 02 11:28:41 crc kubenswrapper[4751]: I1002 11:28:41.217915 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2j96" podStartSLOduration=3.071777936 podStartE2EDuration="9.217894187s" podCreationTimestamp="2025-10-02 11:28:32 +0000 UTC" firstStartedPulling="2025-10-02 11:28:34.093106279 +0000 UTC m=+2196.147332739" lastFinishedPulling="2025-10-02 11:28:40.23922252 +0000 UTC m=+2202.293448990" observedRunningTime="2025-10-02 11:28:41.212656056 +0000 UTC m=+2203.266882526" watchObservedRunningTime="2025-10-02 11:28:41.217894187 +0000 UTC m=+2203.272120637" Oct 02 11:28:42 crc kubenswrapper[4751]: I1002 11:28:42.675390 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:42 crc kubenswrapper[4751]: I1002 11:28:42.675777 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:43 crc kubenswrapper[4751]: I1002 11:28:43.720578 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z2j96" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="registry-server" probeResult="failure" output=< Oct 02 11:28:43 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 11:28:43 crc kubenswrapper[4751]: > Oct 02 11:28:47 crc kubenswrapper[4751]: I1002 11:28:47.228249 4751 generic.go:334] "Generic (PLEG): container finished" podID="74695503-3525-42a8-ad11-0e6644eebcee" containerID="0c0e3f264a78c9516723637f06a4f974974012d6804d5b8d031fc1efa7b0d0ae" exitCode=0 Oct 02 11:28:47 crc kubenswrapper[4751]: I1002 11:28:47.228629 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" event={"ID":"74695503-3525-42a8-ad11-0e6644eebcee","Type":"ContainerDied","Data":"0c0e3f264a78c9516723637f06a4f974974012d6804d5b8d031fc1efa7b0d0ae"} Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.048404 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rtkfm"] Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.076855 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rtkfm"] Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.659572 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.710121 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx\") pod \"74695503-3525-42a8-ad11-0e6644eebcee\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.710338 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key\") pod \"74695503-3525-42a8-ad11-0e6644eebcee\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.710498 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory\") pod \"74695503-3525-42a8-ad11-0e6644eebcee\" (UID: \"74695503-3525-42a8-ad11-0e6644eebcee\") " Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.717865 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx" (OuterVolumeSpecName: "kube-api-access-766fx") pod "74695503-3525-42a8-ad11-0e6644eebcee" (UID: "74695503-3525-42a8-ad11-0e6644eebcee"). InnerVolumeSpecName "kube-api-access-766fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.738566 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74695503-3525-42a8-ad11-0e6644eebcee" (UID: "74695503-3525-42a8-ad11-0e6644eebcee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.738795 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory" (OuterVolumeSpecName: "inventory") pod "74695503-3525-42a8-ad11-0e6644eebcee" (UID: "74695503-3525-42a8-ad11-0e6644eebcee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.813032 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-766fx\" (UniqueName: \"kubernetes.io/projected/74695503-3525-42a8-ad11-0e6644eebcee-kube-api-access-766fx\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.813079 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:48 crc kubenswrapper[4751]: I1002 11:28:48.813092 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74695503-3525-42a8-ad11-0e6644eebcee-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.249555 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" event={"ID":"74695503-3525-42a8-ad11-0e6644eebcee","Type":"ContainerDied","Data":"48fbccdd79da499b5fcbb50a95636be5c71b0933a6ee8258b9af897041bebb81"} Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.249603 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48fbccdd79da499b5fcbb50a95636be5c71b0933a6ee8258b9af897041bebb81" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.249673 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.318080 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9"] Oct 02 11:28:49 crc kubenswrapper[4751]: E1002 11:28:49.318490 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74695503-3525-42a8-ad11-0e6644eebcee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.318512 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="74695503-3525-42a8-ad11-0e6644eebcee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.318739 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="74695503-3525-42a8-ad11-0e6644eebcee" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.319440 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.321724 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.322237 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.325636 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.328276 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.334025 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9"] Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.422399 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.422551 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.422583 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g58sl\" (UniqueName: \"kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.524863 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g58sl\" (UniqueName: \"kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.525097 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.525304 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.530002 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.530123 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.540552 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g58sl\" (UniqueName: \"kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ntcj9\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.560668 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f387d75d-a7dc-4ca1-8a55-d24502420fbb" path="/var/lib/kubelet/pods/f387d75d-a7dc-4ca1-8a55-d24502420fbb/volumes" Oct 02 11:28:49 crc kubenswrapper[4751]: I1002 11:28:49.646202 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:28:50 crc kubenswrapper[4751]: I1002 11:28:50.174509 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9"] Oct 02 11:28:50 crc kubenswrapper[4751]: W1002 11:28:50.181018 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3f64ee3_4f5a_4133_8c31_7f8a770623ee.slice/crio-1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2 WatchSource:0}: Error finding container 1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2: Status 404 returned error can't find the container with id 1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2 Oct 02 11:28:50 crc kubenswrapper[4751]: I1002 11:28:50.260072 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" event={"ID":"a3f64ee3-4f5a-4133-8c31-7f8a770623ee","Type":"ContainerStarted","Data":"1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2"} Oct 02 11:28:52 crc kubenswrapper[4751]: I1002 11:28:52.724162 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:52 crc kubenswrapper[4751]: I1002 11:28:52.779295 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:52 crc kubenswrapper[4751]: I1002 11:28:52.959428 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:53 crc kubenswrapper[4751]: I1002 11:28:53.297271 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" event={"ID":"a3f64ee3-4f5a-4133-8c31-7f8a770623ee","Type":"ContainerStarted","Data":"d7569b433c17cc058017f50c631bcf761bf387f8af95a90a255dfaa8eadd93e0"} Oct 02 11:28:53 crc kubenswrapper[4751]: I1002 11:28:53.319554 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" podStartSLOduration=1.669649931 podStartE2EDuration="4.319532489s" podCreationTimestamp="2025-10-02 11:28:49 +0000 UTC" firstStartedPulling="2025-10-02 11:28:50.184242082 +0000 UTC m=+2212.238468532" lastFinishedPulling="2025-10-02 11:28:52.83412464 +0000 UTC m=+2214.888351090" observedRunningTime="2025-10-02 11:28:53.314267657 +0000 UTC m=+2215.368494127" watchObservedRunningTime="2025-10-02 11:28:53.319532489 +0000 UTC m=+2215.373758939" Oct 02 11:28:54 crc kubenswrapper[4751]: I1002 11:28:54.308618 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2j96" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="registry-server" containerID="cri-o://e9caf54f7a4791b32a6cb5e4f3136c74c254fa97baab118606806d9b85016594" gracePeriod=2 Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.317555 4751 generic.go:334] "Generic (PLEG): container finished" podID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerID="e9caf54f7a4791b32a6cb5e4f3136c74c254fa97baab118606806d9b85016594" exitCode=0 Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.317642 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerDied","Data":"e9caf54f7a4791b32a6cb5e4f3136c74c254fa97baab118606806d9b85016594"} Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.317878 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2j96" event={"ID":"581e8e99-9f68-4ce1-8324-2dab841c957b","Type":"ContainerDied","Data":"b2d3404c58d9974edd2fe32b906496c9a7884201d2d5dcab8829df6abd78a3a7"} Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.317898 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d3404c58d9974edd2fe32b906496c9a7884201d2d5dcab8829df6abd78a3a7" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.390279 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.436573 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content\") pod \"581e8e99-9f68-4ce1-8324-2dab841c957b\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.436685 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvtxg\" (UniqueName: \"kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg\") pod \"581e8e99-9f68-4ce1-8324-2dab841c957b\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.436925 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities\") pod \"581e8e99-9f68-4ce1-8324-2dab841c957b\" (UID: \"581e8e99-9f68-4ce1-8324-2dab841c957b\") " Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.438141 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities" (OuterVolumeSpecName: "utilities") pod "581e8e99-9f68-4ce1-8324-2dab841c957b" (UID: "581e8e99-9f68-4ce1-8324-2dab841c957b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.442932 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg" (OuterVolumeSpecName: "kube-api-access-hvtxg") pod "581e8e99-9f68-4ce1-8324-2dab841c957b" (UID: "581e8e99-9f68-4ce1-8324-2dab841c957b"). InnerVolumeSpecName "kube-api-access-hvtxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.514249 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "581e8e99-9f68-4ce1-8324-2dab841c957b" (UID: "581e8e99-9f68-4ce1-8324-2dab841c957b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.538760 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.538788 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvtxg\" (UniqueName: \"kubernetes.io/projected/581e8e99-9f68-4ce1-8324-2dab841c957b-kube-api-access-hvtxg\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:55 crc kubenswrapper[4751]: I1002 11:28:55.538799 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/581e8e99-9f68-4ce1-8324-2dab841c957b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:28:56 crc kubenswrapper[4751]: I1002 11:28:56.329807 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2j96" Oct 02 11:28:56 crc kubenswrapper[4751]: I1002 11:28:56.353624 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:56 crc kubenswrapper[4751]: I1002 11:28:56.361387 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2j96"] Oct 02 11:28:57 crc kubenswrapper[4751]: I1002 11:28:57.560866 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" path="/var/lib/kubelet/pods/581e8e99-9f68-4ce1-8324-2dab841c957b/volumes" Oct 02 11:29:01 crc kubenswrapper[4751]: I1002 11:29:01.507618 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:29:01 crc kubenswrapper[4751]: I1002 11:29:01.508177 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.067951 4751 scope.go:117] "RemoveContainer" containerID="d961c7bdb06f272c351e6bc2065dc3bd357371cebf470fa61108f7b01d736514" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.119106 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:11 crc kubenswrapper[4751]: E1002 11:29:11.119502 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="extract-content" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.119524 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="extract-content" Oct 02 11:29:11 crc kubenswrapper[4751]: E1002 11:29:11.119545 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="registry-server" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.119553 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="registry-server" Oct 02 11:29:11 crc kubenswrapper[4751]: E1002 11:29:11.119565 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="extract-utilities" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.119573 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="extract-utilities" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.119765 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="581e8e99-9f68-4ce1-8324-2dab841c957b" containerName="registry-server" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.121098 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.130392 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xmxn\" (UniqueName: \"kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.130568 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.130603 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.133375 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.232773 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.233378 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.233577 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.233633 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.233769 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xmxn\" (UniqueName: \"kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.251125 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xmxn\" (UniqueName: \"kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn\") pod \"redhat-marketplace-bnhsk\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.477359 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:11 crc kubenswrapper[4751]: I1002 11:29:11.915235 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:12 crc kubenswrapper[4751]: I1002 11:29:12.465817 4751 generic.go:334] "Generic (PLEG): container finished" podID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerID="0560767d7e605f16550eb78717cdc59bd01ce9dfc8da4c6256ee66b70b936533" exitCode=0 Oct 02 11:29:12 crc kubenswrapper[4751]: I1002 11:29:12.466382 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerDied","Data":"0560767d7e605f16550eb78717cdc59bd01ce9dfc8da4c6256ee66b70b936533"} Oct 02 11:29:12 crc kubenswrapper[4751]: I1002 11:29:12.466416 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerStarted","Data":"5521747077cbf034668294281f7a504fc97e92696e7937e2d455588357242d19"} Oct 02 11:29:16 crc kubenswrapper[4751]: I1002 11:29:16.501015 4751 generic.go:334] "Generic (PLEG): container finished" podID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerID="2e11da9f820f4b9c1523ed2ccbfb4d0ce8ac9359a4b8ecc57884a051ed501035" exitCode=0 Oct 02 11:29:16 crc kubenswrapper[4751]: I1002 11:29:16.501089 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerDied","Data":"2e11da9f820f4b9c1523ed2ccbfb4d0ce8ac9359a4b8ecc57884a051ed501035"} Oct 02 11:29:19 crc kubenswrapper[4751]: I1002 11:29:19.526641 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerStarted","Data":"001468ed5d75391496723e22d5851abe4e43617192d138221cc543ab9ba78d41"} Oct 02 11:29:19 crc kubenswrapper[4751]: I1002 11:29:19.545068 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bnhsk" podStartSLOduration=3.723553421 podStartE2EDuration="8.545053352s" podCreationTimestamp="2025-10-02 11:29:11 +0000 UTC" firstStartedPulling="2025-10-02 11:29:13.478550459 +0000 UTC m=+2235.532776909" lastFinishedPulling="2025-10-02 11:29:18.30005039 +0000 UTC m=+2240.354276840" observedRunningTime="2025-10-02 11:29:19.544041355 +0000 UTC m=+2241.598267825" watchObservedRunningTime="2025-10-02 11:29:19.545053352 +0000 UTC m=+2241.599279802" Oct 02 11:29:21 crc kubenswrapper[4751]: I1002 11:29:21.477447 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:21 crc kubenswrapper[4751]: I1002 11:29:21.477528 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:21 crc kubenswrapper[4751]: I1002 11:29:21.524097 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:31 crc kubenswrapper[4751]: I1002 11:29:31.507594 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:29:31 crc kubenswrapper[4751]: I1002 11:29:31.508122 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:29:31 crc kubenswrapper[4751]: I1002 11:29:31.526827 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:31 crc kubenswrapper[4751]: I1002 11:29:31.574540 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:31 crc kubenswrapper[4751]: I1002 11:29:31.627455 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bnhsk" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="registry-server" containerID="cri-o://001468ed5d75391496723e22d5851abe4e43617192d138221cc543ab9ba78d41" gracePeriod=2 Oct 02 11:29:32 crc kubenswrapper[4751]: I1002 11:29:32.638937 4751 generic.go:334] "Generic (PLEG): container finished" podID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerID="001468ed5d75391496723e22d5851abe4e43617192d138221cc543ab9ba78d41" exitCode=0 Oct 02 11:29:32 crc kubenswrapper[4751]: I1002 11:29:32.639008 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerDied","Data":"001468ed5d75391496723e22d5851abe4e43617192d138221cc543ab9ba78d41"} Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.364503 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.547888 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content\") pod \"84f575f1-eca1-4a74-89e2-a5541bf810c3\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.547949 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities\") pod \"84f575f1-eca1-4a74-89e2-a5541bf810c3\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.548149 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xmxn\" (UniqueName: \"kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn\") pod \"84f575f1-eca1-4a74-89e2-a5541bf810c3\" (UID: \"84f575f1-eca1-4a74-89e2-a5541bf810c3\") " Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.548909 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities" (OuterVolumeSpecName: "utilities") pod "84f575f1-eca1-4a74-89e2-a5541bf810c3" (UID: "84f575f1-eca1-4a74-89e2-a5541bf810c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.560942 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn" (OuterVolumeSpecName: "kube-api-access-2xmxn") pod "84f575f1-eca1-4a74-89e2-a5541bf810c3" (UID: "84f575f1-eca1-4a74-89e2-a5541bf810c3"). InnerVolumeSpecName "kube-api-access-2xmxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.569826 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84f575f1-eca1-4a74-89e2-a5541bf810c3" (UID: "84f575f1-eca1-4a74-89e2-a5541bf810c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.651895 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.652052 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xmxn\" (UniqueName: \"kubernetes.io/projected/84f575f1-eca1-4a74-89e2-a5541bf810c3-kube-api-access-2xmxn\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.652275 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f575f1-eca1-4a74-89e2-a5541bf810c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.656035 4751 generic.go:334] "Generic (PLEG): container finished" podID="a3f64ee3-4f5a-4133-8c31-7f8a770623ee" containerID="d7569b433c17cc058017f50c631bcf761bf387f8af95a90a255dfaa8eadd93e0" exitCode=0 Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.656124 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" event={"ID":"a3f64ee3-4f5a-4133-8c31-7f8a770623ee","Type":"ContainerDied","Data":"d7569b433c17cc058017f50c631bcf761bf387f8af95a90a255dfaa8eadd93e0"} Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.660310 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnhsk" event={"ID":"84f575f1-eca1-4a74-89e2-a5541bf810c3","Type":"ContainerDied","Data":"5521747077cbf034668294281f7a504fc97e92696e7937e2d455588357242d19"} Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.660369 4751 scope.go:117] "RemoveContainer" containerID="001468ed5d75391496723e22d5851abe4e43617192d138221cc543ab9ba78d41" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.660425 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnhsk" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.686325 4751 scope.go:117] "RemoveContainer" containerID="2e11da9f820f4b9c1523ed2ccbfb4d0ce8ac9359a4b8ecc57884a051ed501035" Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.698998 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.705087 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnhsk"] Oct 02 11:29:33 crc kubenswrapper[4751]: I1002 11:29:33.722013 4751 scope.go:117] "RemoveContainer" containerID="0560767d7e605f16550eb78717cdc59bd01ce9dfc8da4c6256ee66b70b936533" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.064526 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.177773 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key\") pod \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.178049 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g58sl\" (UniqueName: \"kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl\") pod \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.178095 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory\") pod \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\" (UID: \"a3f64ee3-4f5a-4133-8c31-7f8a770623ee\") " Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.183391 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl" (OuterVolumeSpecName: "kube-api-access-g58sl") pod "a3f64ee3-4f5a-4133-8c31-7f8a770623ee" (UID: "a3f64ee3-4f5a-4133-8c31-7f8a770623ee"). InnerVolumeSpecName "kube-api-access-g58sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.210241 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory" (OuterVolumeSpecName: "inventory") pod "a3f64ee3-4f5a-4133-8c31-7f8a770623ee" (UID: "a3f64ee3-4f5a-4133-8c31-7f8a770623ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.210266 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a3f64ee3-4f5a-4133-8c31-7f8a770623ee" (UID: "a3f64ee3-4f5a-4133-8c31-7f8a770623ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.280153 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g58sl\" (UniqueName: \"kubernetes.io/projected/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-kube-api-access-g58sl\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.280195 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.280205 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a3f64ee3-4f5a-4133-8c31-7f8a770623ee-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.561852 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" path="/var/lib/kubelet/pods/84f575f1-eca1-4a74-89e2-a5541bf810c3/volumes" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.681714 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" event={"ID":"a3f64ee3-4f5a-4133-8c31-7f8a770623ee","Type":"ContainerDied","Data":"1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2"} Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.681759 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d73ce5ae603341ea43b81be3b7a5adb3de1fabb2a7032f0fb07588c0606b6c2" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.681772 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ntcj9" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.743796 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk"] Oct 02 11:29:35 crc kubenswrapper[4751]: E1002 11:29:35.744245 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="extract-content" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744263 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="extract-content" Oct 02 11:29:35 crc kubenswrapper[4751]: E1002 11:29:35.744279 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="extract-utilities" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744286 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="extract-utilities" Oct 02 11:29:35 crc kubenswrapper[4751]: E1002 11:29:35.744301 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="registry-server" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744310 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="registry-server" Oct 02 11:29:35 crc kubenswrapper[4751]: E1002 11:29:35.744332 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f64ee3-4f5a-4133-8c31-7f8a770623ee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744339 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f64ee3-4f5a-4133-8c31-7f8a770623ee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744572 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f64ee3-4f5a-4133-8c31-7f8a770623ee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.744597 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f575f1-eca1-4a74-89e2-a5541bf810c3" containerName="registry-server" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.745325 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.747901 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.747900 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.747954 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.748274 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.756929 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk"] Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.889208 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfxdl\" (UniqueName: \"kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.889572 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.889754 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.991241 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.991371 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.991413 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfxdl\" (UniqueName: \"kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.995212 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:35 crc kubenswrapper[4751]: I1002 11:29:35.996329 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:36 crc kubenswrapper[4751]: I1002 11:29:36.025683 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfxdl\" (UniqueName: \"kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:36 crc kubenswrapper[4751]: I1002 11:29:36.072790 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:29:36 crc kubenswrapper[4751]: I1002 11:29:36.564956 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk"] Oct 02 11:29:36 crc kubenswrapper[4751]: I1002 11:29:36.690599 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" event={"ID":"0b5f4bef-7365-4e42-8eb4-2165c653e49a","Type":"ContainerStarted","Data":"4eadb1df1a380604502605cdafb99fbfed71fe86d0770b6860c35d8efb926b87"} Oct 02 11:29:38 crc kubenswrapper[4751]: I1002 11:29:38.709937 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" event={"ID":"0b5f4bef-7365-4e42-8eb4-2165c653e49a","Type":"ContainerStarted","Data":"2070b0e22f09b4c1f1484ec428b3e0de172a5a3b90d73e2dfc9b5cc166a87d13"} Oct 02 11:29:39 crc kubenswrapper[4751]: I1002 11:29:39.736120 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" podStartSLOduration=3.264424152 podStartE2EDuration="4.736100403s" podCreationTimestamp="2025-10-02 11:29:35 +0000 UTC" firstStartedPulling="2025-10-02 11:29:36.574986533 +0000 UTC m=+2258.629213003" lastFinishedPulling="2025-10-02 11:29:38.046662804 +0000 UTC m=+2260.100889254" observedRunningTime="2025-10-02 11:29:39.733596666 +0000 UTC m=+2261.787823116" watchObservedRunningTime="2025-10-02 11:29:39.736100403 +0000 UTC m=+2261.790326863" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.143915 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9"] Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.145915 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.147919 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.148214 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.159305 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9"] Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.248720 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.248865 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps9bz\" (UniqueName: \"kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.248926 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.350884 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps9bz\" (UniqueName: \"kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.351009 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.351089 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.352056 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.357520 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.366625 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps9bz\" (UniqueName: \"kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz\") pod \"collect-profiles-29323410-5zjb9\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.476292 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:00 crc kubenswrapper[4751]: I1002 11:30:00.940386 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9"] Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.507414 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.508246 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.508329 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.509390 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.509466 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" gracePeriod=600 Oct 02 11:30:01 crc kubenswrapper[4751]: E1002 11:30:01.639235 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.906300 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" exitCode=0 Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.906406 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f"} Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.906465 4751 scope.go:117] "RemoveContainer" containerID="1e060a3308b00bd54c0249f219d66cb336858bd0585649405985fb25caab9d34" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.907140 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:30:01 crc kubenswrapper[4751]: E1002 11:30:01.907475 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.909232 4751 generic.go:334] "Generic (PLEG): container finished" podID="7456eeb2-be8e-4c1e-a175-ea64db231d46" containerID="4890ecf1811e2780779ec180940eb2c29b71e6751777e974ef96dafb26cb0d89" exitCode=0 Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.909285 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" event={"ID":"7456eeb2-be8e-4c1e-a175-ea64db231d46","Type":"ContainerDied","Data":"4890ecf1811e2780779ec180940eb2c29b71e6751777e974ef96dafb26cb0d89"} Oct 02 11:30:01 crc kubenswrapper[4751]: I1002 11:30:01.909332 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" event={"ID":"7456eeb2-be8e-4c1e-a175-ea64db231d46","Type":"ContainerStarted","Data":"6df1e238000a09d2f446436e4ee7979717dee8e0cf8df50e547583055376102d"} Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.238342 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.414122 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps9bz\" (UniqueName: \"kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz\") pod \"7456eeb2-be8e-4c1e-a175-ea64db231d46\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.414221 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume\") pod \"7456eeb2-be8e-4c1e-a175-ea64db231d46\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.414411 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume\") pod \"7456eeb2-be8e-4c1e-a175-ea64db231d46\" (UID: \"7456eeb2-be8e-4c1e-a175-ea64db231d46\") " Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.415463 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume" (OuterVolumeSpecName: "config-volume") pod "7456eeb2-be8e-4c1e-a175-ea64db231d46" (UID: "7456eeb2-be8e-4c1e-a175-ea64db231d46"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.425032 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz" (OuterVolumeSpecName: "kube-api-access-ps9bz") pod "7456eeb2-be8e-4c1e-a175-ea64db231d46" (UID: "7456eeb2-be8e-4c1e-a175-ea64db231d46"). InnerVolumeSpecName "kube-api-access-ps9bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.437812 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7456eeb2-be8e-4c1e-a175-ea64db231d46" (UID: "7456eeb2-be8e-4c1e-a175-ea64db231d46"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.520854 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7456eeb2-be8e-4c1e-a175-ea64db231d46-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.520933 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps9bz\" (UniqueName: \"kubernetes.io/projected/7456eeb2-be8e-4c1e-a175-ea64db231d46-kube-api-access-ps9bz\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.520968 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7456eeb2-be8e-4c1e-a175-ea64db231d46-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.931476 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" event={"ID":"7456eeb2-be8e-4c1e-a175-ea64db231d46","Type":"ContainerDied","Data":"6df1e238000a09d2f446436e4ee7979717dee8e0cf8df50e547583055376102d"} Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.931527 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6df1e238000a09d2f446436e4ee7979717dee8e0cf8df50e547583055376102d" Oct 02 11:30:03 crc kubenswrapper[4751]: I1002 11:30:03.931591 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9" Oct 02 11:30:04 crc kubenswrapper[4751]: I1002 11:30:04.310763 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw"] Oct 02 11:30:04 crc kubenswrapper[4751]: I1002 11:30:04.319427 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-wt5rw"] Oct 02 11:30:05 crc kubenswrapper[4751]: I1002 11:30:05.562679 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8d6102-cb02-4184-9ad5-e1fa0a15b406" path="/var/lib/kubelet/pods/3a8d6102-cb02-4184-9ad5-e1fa0a15b406/volumes" Oct 02 11:30:11 crc kubenswrapper[4751]: I1002 11:30:11.230750 4751 scope.go:117] "RemoveContainer" containerID="a5f1a2e8ff506be9ebabee4247602b35369362cb75006f63c2476fc3d6f99f78" Oct 02 11:30:16 crc kubenswrapper[4751]: I1002 11:30:16.550684 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:30:16 crc kubenswrapper[4751]: E1002 11:30:16.551446 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:30:27 crc kubenswrapper[4751]: I1002 11:30:27.550380 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:30:27 crc kubenswrapper[4751]: E1002 11:30:27.551088 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:30:35 crc kubenswrapper[4751]: I1002 11:30:35.196701 4751 generic.go:334] "Generic (PLEG): container finished" podID="0b5f4bef-7365-4e42-8eb4-2165c653e49a" containerID="2070b0e22f09b4c1f1484ec428b3e0de172a5a3b90d73e2dfc9b5cc166a87d13" exitCode=2 Oct 02 11:30:35 crc kubenswrapper[4751]: I1002 11:30:35.196750 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" event={"ID":"0b5f4bef-7365-4e42-8eb4-2165c653e49a","Type":"ContainerDied","Data":"2070b0e22f09b4c1f1484ec428b3e0de172a5a3b90d73e2dfc9b5cc166a87d13"} Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.607448 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.735153 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key\") pod \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.735834 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory\") pod \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.735868 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfxdl\" (UniqueName: \"kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl\") pod \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\" (UID: \"0b5f4bef-7365-4e42-8eb4-2165c653e49a\") " Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.742788 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl" (OuterVolumeSpecName: "kube-api-access-dfxdl") pod "0b5f4bef-7365-4e42-8eb4-2165c653e49a" (UID: "0b5f4bef-7365-4e42-8eb4-2165c653e49a"). InnerVolumeSpecName "kube-api-access-dfxdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.764201 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory" (OuterVolumeSpecName: "inventory") pod "0b5f4bef-7365-4e42-8eb4-2165c653e49a" (UID: "0b5f4bef-7365-4e42-8eb4-2165c653e49a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.764630 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b5f4bef-7365-4e42-8eb4-2165c653e49a" (UID: "0b5f4bef-7365-4e42-8eb4-2165c653e49a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.838077 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.838127 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfxdl\" (UniqueName: \"kubernetes.io/projected/0b5f4bef-7365-4e42-8eb4-2165c653e49a-kube-api-access-dfxdl\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:36 crc kubenswrapper[4751]: I1002 11:30:36.838140 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b5f4bef-7365-4e42-8eb4-2165c653e49a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:30:37 crc kubenswrapper[4751]: I1002 11:30:37.218076 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" event={"ID":"0b5f4bef-7365-4e42-8eb4-2165c653e49a","Type":"ContainerDied","Data":"4eadb1df1a380604502605cdafb99fbfed71fe86d0770b6860c35d8efb926b87"} Oct 02 11:30:37 crc kubenswrapper[4751]: I1002 11:30:37.218141 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eadb1df1a380604502605cdafb99fbfed71fe86d0770b6860c35d8efb926b87" Oct 02 11:30:37 crc kubenswrapper[4751]: I1002 11:30:37.218141 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk" Oct 02 11:30:39 crc kubenswrapper[4751]: I1002 11:30:39.557566 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:30:39 crc kubenswrapper[4751]: E1002 11:30:39.558311 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.029913 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc"] Oct 02 11:30:44 crc kubenswrapper[4751]: E1002 11:30:44.032671 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7456eeb2-be8e-4c1e-a175-ea64db231d46" containerName="collect-profiles" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.032694 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7456eeb2-be8e-4c1e-a175-ea64db231d46" containerName="collect-profiles" Oct 02 11:30:44 crc kubenswrapper[4751]: E1002 11:30:44.032708 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5f4bef-7365-4e42-8eb4-2165c653e49a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.032718 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5f4bef-7365-4e42-8eb4-2165c653e49a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.032956 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7456eeb2-be8e-4c1e-a175-ea64db231d46" containerName="collect-profiles" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.032994 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b5f4bef-7365-4e42-8eb4-2165c653e49a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.034543 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.036233 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.036493 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.036515 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.037912 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.045547 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc"] Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.180361 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.180407 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.180574 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg8dh\" (UniqueName: \"kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.283283 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.283331 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.283431 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg8dh\" (UniqueName: \"kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.291095 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.291145 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.305979 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg8dh\" (UniqueName: \"kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.358787 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:30:44 crc kubenswrapper[4751]: I1002 11:30:44.882521 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc"] Oct 02 11:30:45 crc kubenswrapper[4751]: I1002 11:30:45.298461 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" event={"ID":"6768867d-b798-4bb1-be1d-d329d01b5a7f","Type":"ContainerStarted","Data":"4345a422253db43065ba2cb8a72b05b36935dcdf5d56fe87d58b4224f9a8d9cb"} Oct 02 11:30:46 crc kubenswrapper[4751]: I1002 11:30:46.307118 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" event={"ID":"6768867d-b798-4bb1-be1d-d329d01b5a7f","Type":"ContainerStarted","Data":"706b218025b6d399296eda235f2023af028446591562ebe9b46fdd662f694c25"} Oct 02 11:30:46 crc kubenswrapper[4751]: I1002 11:30:46.335660 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" podStartSLOduration=1.524678797 podStartE2EDuration="2.33563546s" podCreationTimestamp="2025-10-02 11:30:44 +0000 UTC" firstStartedPulling="2025-10-02 11:30:44.890959586 +0000 UTC m=+2326.945186036" lastFinishedPulling="2025-10-02 11:30:45.701916249 +0000 UTC m=+2327.756142699" observedRunningTime="2025-10-02 11:30:46.324287206 +0000 UTC m=+2328.378513666" watchObservedRunningTime="2025-10-02 11:30:46.33563546 +0000 UTC m=+2328.389861930" Oct 02 11:30:50 crc kubenswrapper[4751]: I1002 11:30:50.550793 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:30:50 crc kubenswrapper[4751]: E1002 11:30:50.551716 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:01 crc kubenswrapper[4751]: I1002 11:31:01.551001 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:31:01 crc kubenswrapper[4751]: E1002 11:31:01.551906 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:16 crc kubenswrapper[4751]: I1002 11:31:16.550760 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:31:16 crc kubenswrapper[4751]: E1002 11:31:16.551576 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:27 crc kubenswrapper[4751]: I1002 11:31:27.550349 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:31:27 crc kubenswrapper[4751]: E1002 11:31:27.551063 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:32 crc kubenswrapper[4751]: I1002 11:31:32.734293 4751 generic.go:334] "Generic (PLEG): container finished" podID="6768867d-b798-4bb1-be1d-d329d01b5a7f" containerID="706b218025b6d399296eda235f2023af028446591562ebe9b46fdd662f694c25" exitCode=0 Oct 02 11:31:32 crc kubenswrapper[4751]: I1002 11:31:32.734389 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" event={"ID":"6768867d-b798-4bb1-be1d-d329d01b5a7f","Type":"ContainerDied","Data":"706b218025b6d399296eda235f2023af028446591562ebe9b46fdd662f694c25"} Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.176231 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.292691 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory\") pod \"6768867d-b798-4bb1-be1d-d329d01b5a7f\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.292756 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key\") pod \"6768867d-b798-4bb1-be1d-d329d01b5a7f\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.292965 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg8dh\" (UniqueName: \"kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh\") pod \"6768867d-b798-4bb1-be1d-d329d01b5a7f\" (UID: \"6768867d-b798-4bb1-be1d-d329d01b5a7f\") " Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.298905 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh" (OuterVolumeSpecName: "kube-api-access-dg8dh") pod "6768867d-b798-4bb1-be1d-d329d01b5a7f" (UID: "6768867d-b798-4bb1-be1d-d329d01b5a7f"). InnerVolumeSpecName "kube-api-access-dg8dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.320595 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6768867d-b798-4bb1-be1d-d329d01b5a7f" (UID: "6768867d-b798-4bb1-be1d-d329d01b5a7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.321672 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory" (OuterVolumeSpecName: "inventory") pod "6768867d-b798-4bb1-be1d-d329d01b5a7f" (UID: "6768867d-b798-4bb1-be1d-d329d01b5a7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.395219 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg8dh\" (UniqueName: \"kubernetes.io/projected/6768867d-b798-4bb1-be1d-d329d01b5a7f-kube-api-access-dg8dh\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.395266 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.395278 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6768867d-b798-4bb1-be1d-d329d01b5a7f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.753823 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" event={"ID":"6768867d-b798-4bb1-be1d-d329d01b5a7f","Type":"ContainerDied","Data":"4345a422253db43065ba2cb8a72b05b36935dcdf5d56fe87d58b4224f9a8d9cb"} Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.753875 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4345a422253db43065ba2cb8a72b05b36935dcdf5d56fe87d58b4224f9a8d9cb" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.753908 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.838054 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-fk6rt"] Oct 02 11:31:34 crc kubenswrapper[4751]: E1002 11:31:34.838808 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6768867d-b798-4bb1-be1d-d329d01b5a7f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.838835 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="6768867d-b798-4bb1-be1d-d329d01b5a7f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.839083 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="6768867d-b798-4bb1-be1d-d329d01b5a7f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.839745 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.841847 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.844621 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-fk6rt"] Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.845060 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.845225 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.845374 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.903884 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.903980 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjjl8\" (UniqueName: \"kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:34 crc kubenswrapper[4751]: I1002 11:31:34.904116 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.005452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjjl8\" (UniqueName: \"kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.005652 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.005859 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.014203 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.014380 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.031806 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjjl8\" (UniqueName: \"kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8\") pod \"ssh-known-hosts-edpm-deployment-fk6rt\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.161318 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.678051 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-fk6rt"] Oct 02 11:31:35 crc kubenswrapper[4751]: I1002 11:31:35.773043 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" event={"ID":"a3bc6e94-6738-40cf-9d23-e14ab0edc63b","Type":"ContainerStarted","Data":"72f4a57ef7e1d656d417b110f66169efbf6248ad6b0a395022eff1ceaaf615fd"} Oct 02 11:31:36 crc kubenswrapper[4751]: I1002 11:31:36.785042 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" event={"ID":"a3bc6e94-6738-40cf-9d23-e14ab0edc63b","Type":"ContainerStarted","Data":"74c43c0579bf7e305fa3ec87791c4b846d72c9ce00c0fe7142a90cc73108847e"} Oct 02 11:31:36 crc kubenswrapper[4751]: I1002 11:31:36.809692 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" podStartSLOduration=2.622685407 podStartE2EDuration="2.809673959s" podCreationTimestamp="2025-10-02 11:31:34 +0000 UTC" firstStartedPulling="2025-10-02 11:31:35.692298256 +0000 UTC m=+2377.746524706" lastFinishedPulling="2025-10-02 11:31:35.879286808 +0000 UTC m=+2377.933513258" observedRunningTime="2025-10-02 11:31:36.80302269 +0000 UTC m=+2378.857249140" watchObservedRunningTime="2025-10-02 11:31:36.809673959 +0000 UTC m=+2378.863900409" Oct 02 11:31:41 crc kubenswrapper[4751]: I1002 11:31:41.549847 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:31:41 crc kubenswrapper[4751]: E1002 11:31:41.550665 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:44 crc kubenswrapper[4751]: I1002 11:31:44.860907 4751 generic.go:334] "Generic (PLEG): container finished" podID="a3bc6e94-6738-40cf-9d23-e14ab0edc63b" containerID="74c43c0579bf7e305fa3ec87791c4b846d72c9ce00c0fe7142a90cc73108847e" exitCode=0 Oct 02 11:31:44 crc kubenswrapper[4751]: I1002 11:31:44.860991 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" event={"ID":"a3bc6e94-6738-40cf-9d23-e14ab0edc63b","Type":"ContainerDied","Data":"74c43c0579bf7e305fa3ec87791c4b846d72c9ce00c0fe7142a90cc73108847e"} Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.248246 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.311765 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam\") pod \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.312067 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjjl8\" (UniqueName: \"kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8\") pod \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.312779 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0\") pod \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\" (UID: \"a3bc6e94-6738-40cf-9d23-e14ab0edc63b\") " Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.317608 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8" (OuterVolumeSpecName: "kube-api-access-sjjl8") pod "a3bc6e94-6738-40cf-9d23-e14ab0edc63b" (UID: "a3bc6e94-6738-40cf-9d23-e14ab0edc63b"). InnerVolumeSpecName "kube-api-access-sjjl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.339769 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a3bc6e94-6738-40cf-9d23-e14ab0edc63b" (UID: "a3bc6e94-6738-40cf-9d23-e14ab0edc63b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.341629 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "a3bc6e94-6738-40cf-9d23-e14ab0edc63b" (UID: "a3bc6e94-6738-40cf-9d23-e14ab0edc63b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.415197 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjjl8\" (UniqueName: \"kubernetes.io/projected/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-kube-api-access-sjjl8\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.415226 4751 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.415235 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a3bc6e94-6738-40cf-9d23-e14ab0edc63b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.880241 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" event={"ID":"a3bc6e94-6738-40cf-9d23-e14ab0edc63b","Type":"ContainerDied","Data":"72f4a57ef7e1d656d417b110f66169efbf6248ad6b0a395022eff1ceaaf615fd"} Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.880565 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72f4a57ef7e1d656d417b110f66169efbf6248ad6b0a395022eff1ceaaf615fd" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.880247 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-fk6rt" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.944119 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q"] Oct 02 11:31:46 crc kubenswrapper[4751]: E1002 11:31:46.944516 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bc6e94-6738-40cf-9d23-e14ab0edc63b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.944532 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bc6e94-6738-40cf-9d23-e14ab0edc63b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.944729 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3bc6e94-6738-40cf-9d23-e14ab0edc63b" containerName="ssh-known-hosts-edpm-deployment" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.945515 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.948894 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.948968 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.951538 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.958103 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:31:46 crc kubenswrapper[4751]: I1002 11:31:46.959605 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q"] Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.025785 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.025846 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.025875 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hbvg\" (UniqueName: \"kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.127446 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.127528 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.127579 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hbvg\" (UniqueName: \"kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.132374 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.132498 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.146145 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hbvg\" (UniqueName: \"kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-fwn4q\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.264612 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.775559 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q"] Oct 02 11:31:47 crc kubenswrapper[4751]: I1002 11:31:47.893885 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" event={"ID":"8dbf0347-7940-48ca-ad8c-29329d4b0391","Type":"ContainerStarted","Data":"02de0d6b1c6ba188620dc5ada194dd516fffa3058af48daac6e51e7ce72ea07a"} Oct 02 11:31:48 crc kubenswrapper[4751]: I1002 11:31:48.905803 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" event={"ID":"8dbf0347-7940-48ca-ad8c-29329d4b0391","Type":"ContainerStarted","Data":"0f6c1840c7e0ae66f0a11c07018f06d54abe6106f5a77d9a0805301608a4f582"} Oct 02 11:31:48 crc kubenswrapper[4751]: I1002 11:31:48.923861 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" podStartSLOduration=2.7753085669999997 podStartE2EDuration="2.923843277s" podCreationTimestamp="2025-10-02 11:31:46 +0000 UTC" firstStartedPulling="2025-10-02 11:31:47.779741585 +0000 UTC m=+2389.833968035" lastFinishedPulling="2025-10-02 11:31:47.928276295 +0000 UTC m=+2389.982502745" observedRunningTime="2025-10-02 11:31:48.920321433 +0000 UTC m=+2390.974547883" watchObservedRunningTime="2025-10-02 11:31:48.923843277 +0000 UTC m=+2390.978069727" Oct 02 11:31:56 crc kubenswrapper[4751]: I1002 11:31:56.550556 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:31:56 crc kubenswrapper[4751]: E1002 11:31:56.551834 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:31:56 crc kubenswrapper[4751]: I1002 11:31:56.974659 4751 generic.go:334] "Generic (PLEG): container finished" podID="8dbf0347-7940-48ca-ad8c-29329d4b0391" containerID="0f6c1840c7e0ae66f0a11c07018f06d54abe6106f5a77d9a0805301608a4f582" exitCode=0 Oct 02 11:31:56 crc kubenswrapper[4751]: I1002 11:31:56.974725 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" event={"ID":"8dbf0347-7940-48ca-ad8c-29329d4b0391","Type":"ContainerDied","Data":"0f6c1840c7e0ae66f0a11c07018f06d54abe6106f5a77d9a0805301608a4f582"} Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.379853 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.434951 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory\") pod \"8dbf0347-7940-48ca-ad8c-29329d4b0391\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.435123 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key\") pod \"8dbf0347-7940-48ca-ad8c-29329d4b0391\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.435284 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hbvg\" (UniqueName: \"kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg\") pod \"8dbf0347-7940-48ca-ad8c-29329d4b0391\" (UID: \"8dbf0347-7940-48ca-ad8c-29329d4b0391\") " Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.440624 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg" (OuterVolumeSpecName: "kube-api-access-9hbvg") pod "8dbf0347-7940-48ca-ad8c-29329d4b0391" (UID: "8dbf0347-7940-48ca-ad8c-29329d4b0391"). InnerVolumeSpecName "kube-api-access-9hbvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.469887 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8dbf0347-7940-48ca-ad8c-29329d4b0391" (UID: "8dbf0347-7940-48ca-ad8c-29329d4b0391"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.470043 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory" (OuterVolumeSpecName: "inventory") pod "8dbf0347-7940-48ca-ad8c-29329d4b0391" (UID: "8dbf0347-7940-48ca-ad8c-29329d4b0391"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.537891 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.538187 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hbvg\" (UniqueName: \"kubernetes.io/projected/8dbf0347-7940-48ca-ad8c-29329d4b0391-kube-api-access-9hbvg\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.538265 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbf0347-7940-48ca-ad8c-29329d4b0391-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.993535 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" event={"ID":"8dbf0347-7940-48ca-ad8c-29329d4b0391","Type":"ContainerDied","Data":"02de0d6b1c6ba188620dc5ada194dd516fffa3058af48daac6e51e7ce72ea07a"} Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.993861 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02de0d6b1c6ba188620dc5ada194dd516fffa3058af48daac6e51e7ce72ea07a" Oct 02 11:31:58 crc kubenswrapper[4751]: I1002 11:31:58.993577 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-fwn4q" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.065352 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs"] Oct 02 11:31:59 crc kubenswrapper[4751]: E1002 11:31:59.065869 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf0347-7940-48ca-ad8c-29329d4b0391" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.065899 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf0347-7940-48ca-ad8c-29329d4b0391" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.066119 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf0347-7940-48ca-ad8c-29329d4b0391" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.066828 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.069025 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.069094 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.070041 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.070403 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.075704 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs"] Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.149943 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.150021 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.150246 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbwx4\" (UniqueName: \"kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.251959 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.252026 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.252062 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbwx4\" (UniqueName: \"kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.256521 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.260393 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.269249 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbwx4\" (UniqueName: \"kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.393262 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.401918 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.891475 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs"] Oct 02 11:31:59 crc kubenswrapper[4751]: I1002 11:31:59.896381 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:32:00 crc kubenswrapper[4751]: I1002 11:32:00.005655 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" event={"ID":"400bc737-3667-4245-a0a3-fa5d2935d967","Type":"ContainerStarted","Data":"9a07d7e1d7f90506926041799230bf73990d7ac6afaf97f3e81a9ef8f9c36f5c"} Oct 02 11:32:00 crc kubenswrapper[4751]: I1002 11:32:00.237899 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:32:01 crc kubenswrapper[4751]: I1002 11:32:01.013779 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" event={"ID":"400bc737-3667-4245-a0a3-fa5d2935d967","Type":"ContainerStarted","Data":"5d69a79cce5d10502e4998dc064a180e3b7b503f5fd367013dfcb609e482b71a"} Oct 02 11:32:01 crc kubenswrapper[4751]: I1002 11:32:01.028802 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" podStartSLOduration=1.689839579 podStartE2EDuration="2.028766077s" podCreationTimestamp="2025-10-02 11:31:59 +0000 UTC" firstStartedPulling="2025-10-02 11:31:59.896098904 +0000 UTC m=+2401.950325354" lastFinishedPulling="2025-10-02 11:32:00.235025402 +0000 UTC m=+2402.289251852" observedRunningTime="2025-10-02 11:32:01.025885819 +0000 UTC m=+2403.080112279" watchObservedRunningTime="2025-10-02 11:32:01.028766077 +0000 UTC m=+2403.082992527" Oct 02 11:32:07 crc kubenswrapper[4751]: I1002 11:32:07.549859 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:32:07 crc kubenswrapper[4751]: E1002 11:32:07.550449 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:32:10 crc kubenswrapper[4751]: I1002 11:32:10.094463 4751 generic.go:334] "Generic (PLEG): container finished" podID="400bc737-3667-4245-a0a3-fa5d2935d967" containerID="5d69a79cce5d10502e4998dc064a180e3b7b503f5fd367013dfcb609e482b71a" exitCode=0 Oct 02 11:32:10 crc kubenswrapper[4751]: I1002 11:32:10.094549 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" event={"ID":"400bc737-3667-4245-a0a3-fa5d2935d967","Type":"ContainerDied","Data":"5d69a79cce5d10502e4998dc064a180e3b7b503f5fd367013dfcb609e482b71a"} Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.467606 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.580530 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key\") pod \"400bc737-3667-4245-a0a3-fa5d2935d967\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.580738 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory\") pod \"400bc737-3667-4245-a0a3-fa5d2935d967\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.580869 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbwx4\" (UniqueName: \"kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4\") pod \"400bc737-3667-4245-a0a3-fa5d2935d967\" (UID: \"400bc737-3667-4245-a0a3-fa5d2935d967\") " Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.594170 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4" (OuterVolumeSpecName: "kube-api-access-dbwx4") pod "400bc737-3667-4245-a0a3-fa5d2935d967" (UID: "400bc737-3667-4245-a0a3-fa5d2935d967"). InnerVolumeSpecName "kube-api-access-dbwx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.608353 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "400bc737-3667-4245-a0a3-fa5d2935d967" (UID: "400bc737-3667-4245-a0a3-fa5d2935d967"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.608704 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory" (OuterVolumeSpecName: "inventory") pod "400bc737-3667-4245-a0a3-fa5d2935d967" (UID: "400bc737-3667-4245-a0a3-fa5d2935d967"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.683492 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.683563 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbwx4\" (UniqueName: \"kubernetes.io/projected/400bc737-3667-4245-a0a3-fa5d2935d967-kube-api-access-dbwx4\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:11 crc kubenswrapper[4751]: I1002 11:32:11.683578 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/400bc737-3667-4245-a0a3-fa5d2935d967-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.117263 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" event={"ID":"400bc737-3667-4245-a0a3-fa5d2935d967","Type":"ContainerDied","Data":"9a07d7e1d7f90506926041799230bf73990d7ac6afaf97f3e81a9ef8f9c36f5c"} Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.117309 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a07d7e1d7f90506926041799230bf73990d7ac6afaf97f3e81a9ef8f9c36f5c" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.117375 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.196386 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk"] Oct 02 11:32:12 crc kubenswrapper[4751]: E1002 11:32:12.197433 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="400bc737-3667-4245-a0a3-fa5d2935d967" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.197468 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="400bc737-3667-4245-a0a3-fa5d2935d967" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.197752 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="400bc737-3667-4245-a0a3-fa5d2935d967" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.198467 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.200860 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.200929 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.200867 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.201099 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.201118 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.201138 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.201710 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.203430 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.224001 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk"] Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294020 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294076 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294269 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294326 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294436 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294488 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294518 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294557 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294584 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8zz9\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294663 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294699 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294723 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294748 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.294806 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.396836 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397164 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397295 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397416 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397525 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397623 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397758 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8zz9\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.397941 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.398043 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.398202 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.398767 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.399019 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.399215 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.399329 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.401644 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.402005 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.402677 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.403307 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.403470 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.403923 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.403930 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.404065 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.404121 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.404425 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.404722 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.407760 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.407847 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.414665 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8zz9\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:12 crc kubenswrapper[4751]: I1002 11:32:12.518818 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:13 crc kubenswrapper[4751]: I1002 11:32:13.036145 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk"] Oct 02 11:32:13 crc kubenswrapper[4751]: I1002 11:32:13.130031 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" event={"ID":"3332ad0f-fbf3-455e-8cb9-892c3cb435c2","Type":"ContainerStarted","Data":"ccfab7261ec8bc312de88717203f54d4bdf21615496f2a70a4be6b0b86b3576f"} Oct 02 11:32:14 crc kubenswrapper[4751]: I1002 11:32:14.141673 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" event={"ID":"3332ad0f-fbf3-455e-8cb9-892c3cb435c2","Type":"ContainerStarted","Data":"398b6d28955506690c280ebf1eceb900df73662f0acf951895dd640814df926c"} Oct 02 11:32:14 crc kubenswrapper[4751]: I1002 11:32:14.177093 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" podStartSLOduration=2.014839948 podStartE2EDuration="2.177074351s" podCreationTimestamp="2025-10-02 11:32:12 +0000 UTC" firstStartedPulling="2025-10-02 11:32:13.041170222 +0000 UTC m=+2415.095396672" lastFinishedPulling="2025-10-02 11:32:13.203404595 +0000 UTC m=+2415.257631075" observedRunningTime="2025-10-02 11:32:14.163841573 +0000 UTC m=+2416.218068043" watchObservedRunningTime="2025-10-02 11:32:14.177074351 +0000 UTC m=+2416.231300801" Oct 02 11:32:22 crc kubenswrapper[4751]: I1002 11:32:22.550374 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:32:22 crc kubenswrapper[4751]: E1002 11:32:22.551189 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:32:36 crc kubenswrapper[4751]: I1002 11:32:36.550298 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:32:36 crc kubenswrapper[4751]: E1002 11:32:36.551214 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:32:50 crc kubenswrapper[4751]: I1002 11:32:50.551082 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:32:50 crc kubenswrapper[4751]: E1002 11:32:50.552063 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:32:55 crc kubenswrapper[4751]: I1002 11:32:55.543823 4751 generic.go:334] "Generic (PLEG): container finished" podID="3332ad0f-fbf3-455e-8cb9-892c3cb435c2" containerID="398b6d28955506690c280ebf1eceb900df73662f0acf951895dd640814df926c" exitCode=0 Oct 02 11:32:55 crc kubenswrapper[4751]: I1002 11:32:55.543940 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" event={"ID":"3332ad0f-fbf3-455e-8cb9-892c3cb435c2","Type":"ContainerDied","Data":"398b6d28955506690c280ebf1eceb900df73662f0acf951895dd640814df926c"} Oct 02 11:32:56 crc kubenswrapper[4751]: I1002 11:32:56.945617 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044381 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044518 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044574 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044603 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044623 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8zz9\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044649 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044687 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044748 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044763 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044797 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044819 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044836 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044862 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.044886 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle\") pod \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\" (UID: \"3332ad0f-fbf3-455e-8cb9-892c3cb435c2\") " Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.050931 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.051423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.051799 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.052023 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.052058 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.052549 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.052852 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.053196 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.053158 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9" (OuterVolumeSpecName: "kube-api-access-m8zz9") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "kube-api-access-m8zz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.053277 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.053627 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.055635 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.074198 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory" (OuterVolumeSpecName: "inventory") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.074525 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3332ad0f-fbf3-455e-8cb9-892c3cb435c2" (UID: "3332ad0f-fbf3-455e-8cb9-892c3cb435c2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147541 4751 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147586 4751 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147597 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147608 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147621 4751 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147631 4751 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147642 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8zz9\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-kube-api-access-m8zz9\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147652 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147660 4751 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147669 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147679 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147689 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147700 4751 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.147709 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3332ad0f-fbf3-455e-8cb9-892c3cb435c2-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.562028 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.562516 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk" event={"ID":"3332ad0f-fbf3-455e-8cb9-892c3cb435c2","Type":"ContainerDied","Data":"ccfab7261ec8bc312de88717203f54d4bdf21615496f2a70a4be6b0b86b3576f"} Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.562555 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccfab7261ec8bc312de88717203f54d4bdf21615496f2a70a4be6b0b86b3576f" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.656882 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq"] Oct 02 11:32:57 crc kubenswrapper[4751]: E1002 11:32:57.657407 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3332ad0f-fbf3-455e-8cb9-892c3cb435c2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.657432 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="3332ad0f-fbf3-455e-8cb9-892c3cb435c2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.657694 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="3332ad0f-fbf3-455e-8cb9-892c3cb435c2" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.658874 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.661015 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.661025 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.661638 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.661720 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.662715 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.679351 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq"] Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.758085 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.758195 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.758385 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.758523 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwjzb\" (UniqueName: \"kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.758582 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.860318 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.860423 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.860462 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.860500 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwjzb\" (UniqueName: \"kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.860525 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.861791 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.864656 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.865019 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.865044 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.878700 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwjzb\" (UniqueName: \"kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8txlq\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:57 crc kubenswrapper[4751]: I1002 11:32:57.979090 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:32:58 crc kubenswrapper[4751]: I1002 11:32:58.457547 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq"] Oct 02 11:32:58 crc kubenswrapper[4751]: I1002 11:32:58.571650 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" event={"ID":"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141","Type":"ContainerStarted","Data":"5c0e2349190dc7c66927df9f62755751c85778500f1bddb8ad1301f4cd064e7c"} Oct 02 11:32:59 crc kubenswrapper[4751]: I1002 11:32:59.586872 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" event={"ID":"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141","Type":"ContainerStarted","Data":"152b952bea8347a8dfe5f6a69ef6f1bfd7ff5165b3cd9857f794076a45e8fbff"} Oct 02 11:32:59 crc kubenswrapper[4751]: I1002 11:32:59.609163 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" podStartSLOduration=2.422703735 podStartE2EDuration="2.609137103s" podCreationTimestamp="2025-10-02 11:32:57 +0000 UTC" firstStartedPulling="2025-10-02 11:32:58.469295887 +0000 UTC m=+2460.523522337" lastFinishedPulling="2025-10-02 11:32:58.655729235 +0000 UTC m=+2460.709955705" observedRunningTime="2025-10-02 11:32:59.60127998 +0000 UTC m=+2461.655506430" watchObservedRunningTime="2025-10-02 11:32:59.609137103 +0000 UTC m=+2461.663363573" Oct 02 11:33:01 crc kubenswrapper[4751]: I1002 11:33:01.549769 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:33:01 crc kubenswrapper[4751]: E1002 11:33:01.550287 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:33:15 crc kubenswrapper[4751]: I1002 11:33:15.550877 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:33:15 crc kubenswrapper[4751]: E1002 11:33:15.551710 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:33:29 crc kubenswrapper[4751]: I1002 11:33:29.556967 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:33:29 crc kubenswrapper[4751]: E1002 11:33:29.557819 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:33:42 crc kubenswrapper[4751]: I1002 11:33:42.550935 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:33:42 crc kubenswrapper[4751]: E1002 11:33:42.551796 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:33:57 crc kubenswrapper[4751]: I1002 11:33:57.070789 4751 generic.go:334] "Generic (PLEG): container finished" podID="bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" containerID="152b952bea8347a8dfe5f6a69ef6f1bfd7ff5165b3cd9857f794076a45e8fbff" exitCode=0 Oct 02 11:33:57 crc kubenswrapper[4751]: I1002 11:33:57.070899 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" event={"ID":"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141","Type":"ContainerDied","Data":"152b952bea8347a8dfe5f6a69ef6f1bfd7ff5165b3cd9857f794076a45e8fbff"} Oct 02 11:33:57 crc kubenswrapper[4751]: I1002 11:33:57.549903 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:33:57 crc kubenswrapper[4751]: E1002 11:33:57.550350 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.461733 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.587825 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle\") pod \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.587918 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwjzb\" (UniqueName: \"kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb\") pod \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.588045 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0\") pod \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.588075 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory\") pod \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.588123 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key\") pod \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\" (UID: \"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141\") " Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.593731 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" (UID: "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.593977 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb" (OuterVolumeSpecName: "kube-api-access-hwjzb") pod "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" (UID: "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141"). InnerVolumeSpecName "kube-api-access-hwjzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.615361 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" (UID: "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.616159 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory" (OuterVolumeSpecName: "inventory") pod "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" (UID: "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.618976 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" (UID: "bf9e4b5e-949a-4a8b-a8fe-dea52bd10141"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.690946 4751 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.691010 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.691025 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.691040 4751 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:58 crc kubenswrapper[4751]: I1002 11:33:58.691052 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwjzb\" (UniqueName: \"kubernetes.io/projected/bf9e4b5e-949a-4a8b-a8fe-dea52bd10141-kube-api-access-hwjzb\") on node \"crc\" DevicePath \"\"" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.088947 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" event={"ID":"bf9e4b5e-949a-4a8b-a8fe-dea52bd10141","Type":"ContainerDied","Data":"5c0e2349190dc7c66927df9f62755751c85778500f1bddb8ad1301f4cd064e7c"} Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.088997 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c0e2349190dc7c66927df9f62755751c85778500f1bddb8ad1301f4cd064e7c" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.089310 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8txlq" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.214700 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf"] Oct 02 11:33:59 crc kubenswrapper[4751]: E1002 11:33:59.215426 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.215461 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.215733 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9e4b5e-949a-4a8b-a8fe-dea52bd10141" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.216857 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.219597 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.220441 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.220511 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.220775 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.220853 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.222114 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.224055 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf"] Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.304839 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.304895 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.304956 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.305013 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp4lg\" (UniqueName: \"kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.305062 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.305100 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.406841 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.407143 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.407269 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.407351 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp4lg\" (UniqueName: \"kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.407446 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.407542 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.409258 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.409526 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.410418 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.410513 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.414627 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.422599 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.422670 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.423017 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.423824 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.425780 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp4lg\" (UniqueName: \"kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.539395 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:33:59 crc kubenswrapper[4751]: I1002 11:33:59.546253 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:34:00 crc kubenswrapper[4751]: I1002 11:34:00.050325 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf"] Oct 02 11:34:00 crc kubenswrapper[4751]: I1002 11:34:00.099523 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" event={"ID":"8cfe9274-53ad-433f-b478-17b0b5b45bff","Type":"ContainerStarted","Data":"4c598244b5f621c8bba2096bd65992f17cae7554f4db8c73459b2c52330be5ab"} Oct 02 11:34:00 crc kubenswrapper[4751]: I1002 11:34:00.279625 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:34:01 crc kubenswrapper[4751]: I1002 11:34:01.109327 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" event={"ID":"8cfe9274-53ad-433f-b478-17b0b5b45bff","Type":"ContainerStarted","Data":"3e8355ce6d8761bd00fd48060a19b7cb2784f17052c439619c14154d1e33eeb4"} Oct 02 11:34:01 crc kubenswrapper[4751]: I1002 11:34:01.135298 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" podStartSLOduration=1.920074632 podStartE2EDuration="2.135280139s" podCreationTimestamp="2025-10-02 11:33:59 +0000 UTC" firstStartedPulling="2025-10-02 11:34:00.061083781 +0000 UTC m=+2522.115310221" lastFinishedPulling="2025-10-02 11:34:00.276289278 +0000 UTC m=+2522.330515728" observedRunningTime="2025-10-02 11:34:01.127792127 +0000 UTC m=+2523.182018587" watchObservedRunningTime="2025-10-02 11:34:01.135280139 +0000 UTC m=+2523.189506589" Oct 02 11:34:12 crc kubenswrapper[4751]: I1002 11:34:12.550061 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:34:12 crc kubenswrapper[4751]: E1002 11:34:12.550883 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:34:26 crc kubenswrapper[4751]: I1002 11:34:26.550129 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:34:26 crc kubenswrapper[4751]: E1002 11:34:26.550873 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:34:41 crc kubenswrapper[4751]: I1002 11:34:41.498799 4751 generic.go:334] "Generic (PLEG): container finished" podID="8cfe9274-53ad-433f-b478-17b0b5b45bff" containerID="3e8355ce6d8761bd00fd48060a19b7cb2784f17052c439619c14154d1e33eeb4" exitCode=0 Oct 02 11:34:41 crc kubenswrapper[4751]: I1002 11:34:41.498884 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" event={"ID":"8cfe9274-53ad-433f-b478-17b0b5b45bff","Type":"ContainerDied","Data":"3e8355ce6d8761bd00fd48060a19b7cb2784f17052c439619c14154d1e33eeb4"} Oct 02 11:34:41 crc kubenswrapper[4751]: I1002 11:34:41.549972 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:34:41 crc kubenswrapper[4751]: E1002 11:34:41.550300 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:34:42 crc kubenswrapper[4751]: I1002 11:34:42.908225 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.035703 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.035774 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.035823 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.036096 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.036226 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp4lg\" (UniqueName: \"kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.036300 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle\") pod \"8cfe9274-53ad-433f-b478-17b0b5b45bff\" (UID: \"8cfe9274-53ad-433f-b478-17b0b5b45bff\") " Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.042151 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.042500 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg" (OuterVolumeSpecName: "kube-api-access-hp4lg") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "kube-api-access-hp4lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.068378 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory" (OuterVolumeSpecName: "inventory") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.077058 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.078525 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.089226 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8cfe9274-53ad-433f-b478-17b0b5b45bff" (UID: "8cfe9274-53ad-433f-b478-17b0b5b45bff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.138427 4751 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.139115 4751 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.139142 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.139152 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.139165 4751 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8cfe9274-53ad-433f-b478-17b0b5b45bff-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.139207 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp4lg\" (UniqueName: \"kubernetes.io/projected/8cfe9274-53ad-433f-b478-17b0b5b45bff-kube-api-access-hp4lg\") on node \"crc\" DevicePath \"\"" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.518014 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" event={"ID":"8cfe9274-53ad-433f-b478-17b0b5b45bff","Type":"ContainerDied","Data":"4c598244b5f621c8bba2096bd65992f17cae7554f4db8c73459b2c52330be5ab"} Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.518048 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c598244b5f621c8bba2096bd65992f17cae7554f4db8c73459b2c52330be5ab" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.518081 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.619244 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9"] Oct 02 11:34:43 crc kubenswrapper[4751]: E1002 11:34:43.619989 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cfe9274-53ad-433f-b478-17b0b5b45bff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.620010 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cfe9274-53ad-433f-b478-17b0b5b45bff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.620323 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cfe9274-53ad-433f-b478-17b0b5b45bff" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.621083 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.622656 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.622898 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.623375 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.623669 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.624385 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.631661 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9"] Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.650323 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.650548 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.650600 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.650672 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.650702 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6cpj\" (UniqueName: \"kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.752333 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.752424 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.752518 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.752555 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6cpj\" (UniqueName: \"kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.752631 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.758705 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.758804 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.758782 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.759409 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.776422 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6cpj\" (UniqueName: \"kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:43 crc kubenswrapper[4751]: I1002 11:34:43.939162 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:34:44 crc kubenswrapper[4751]: I1002 11:34:44.438810 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9"] Oct 02 11:34:44 crc kubenswrapper[4751]: W1002 11:34:44.443599 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef947945_9630_4475_84ea_0798e9707829.slice/crio-2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54 WatchSource:0}: Error finding container 2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54: Status 404 returned error can't find the container with id 2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54 Oct 02 11:34:44 crc kubenswrapper[4751]: I1002 11:34:44.526403 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" event={"ID":"ef947945-9630-4475-84ea-0798e9707829","Type":"ContainerStarted","Data":"2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54"} Oct 02 11:34:45 crc kubenswrapper[4751]: I1002 11:34:45.537097 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" event={"ID":"ef947945-9630-4475-84ea-0798e9707829","Type":"ContainerStarted","Data":"f641763d027dab4e2dc7bf127b27e279b5fadc2bb845159a7d308b65635c65f8"} Oct 02 11:34:45 crc kubenswrapper[4751]: I1002 11:34:45.563557 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" podStartSLOduration=2.3428794059999998 podStartE2EDuration="2.563518821s" podCreationTimestamp="2025-10-02 11:34:43 +0000 UTC" firstStartedPulling="2025-10-02 11:34:44.4464045 +0000 UTC m=+2566.500630950" lastFinishedPulling="2025-10-02 11:34:44.667043915 +0000 UTC m=+2566.721270365" observedRunningTime="2025-10-02 11:34:45.55462382 +0000 UTC m=+2567.608850270" watchObservedRunningTime="2025-10-02 11:34:45.563518821 +0000 UTC m=+2567.617745281" Oct 02 11:34:52 crc kubenswrapper[4751]: I1002 11:34:52.550059 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:34:52 crc kubenswrapper[4751]: E1002 11:34:52.551263 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:35:03 crc kubenswrapper[4751]: I1002 11:35:03.550315 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:35:04 crc kubenswrapper[4751]: I1002 11:35:04.719782 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676"} Oct 02 11:35:11 crc kubenswrapper[4751]: I1002 11:35:11.395655 4751 scope.go:117] "RemoveContainer" containerID="86e2f9486e561822c75867e8ac75be15fa21cb27b9103f872c5ad116d7633e5b" Oct 02 11:35:11 crc kubenswrapper[4751]: I1002 11:35:11.422073 4751 scope.go:117] "RemoveContainer" containerID="142977f5072151cd599716360061c74b96a5a3f377869230e501df8882e47ec7" Oct 02 11:35:11 crc kubenswrapper[4751]: I1002 11:35:11.470221 4751 scope.go:117] "RemoveContainer" containerID="e9caf54f7a4791b32a6cb5e4f3136c74c254fa97baab118606806d9b85016594" Oct 02 11:37:31 crc kubenswrapper[4751]: I1002 11:37:31.507232 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:37:31 crc kubenswrapper[4751]: I1002 11:37:31.507837 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:38:01 crc kubenswrapper[4751]: I1002 11:38:01.507090 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:38:01 crc kubenswrapper[4751]: I1002 11:38:01.507885 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.875421 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.878429 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.900966 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.922728 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.923023 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb62g\" (UniqueName: \"kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:12 crc kubenswrapper[4751]: I1002 11:38:12.923247 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.025399 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.025940 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.026023 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.026229 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb62g\" (UniqueName: \"kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.026318 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.053529 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb62g\" (UniqueName: \"kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g\") pod \"certified-operators-89sqb\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.199473 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:13 crc kubenswrapper[4751]: I1002 11:38:13.500723 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:14 crc kubenswrapper[4751]: I1002 11:38:14.433481 4751 generic.go:334] "Generic (PLEG): container finished" podID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerID="164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6" exitCode=0 Oct 02 11:38:14 crc kubenswrapper[4751]: I1002 11:38:14.433578 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerDied","Data":"164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6"} Oct 02 11:38:14 crc kubenswrapper[4751]: I1002 11:38:14.433850 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerStarted","Data":"558e172ee757ebc6b1444bfaf93d16a67a1461ec97d9016150e2d613c7ae22cb"} Oct 02 11:38:14 crc kubenswrapper[4751]: I1002 11:38:14.435558 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:38:16 crc kubenswrapper[4751]: I1002 11:38:16.457491 4751 generic.go:334] "Generic (PLEG): container finished" podID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerID="c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1" exitCode=0 Oct 02 11:38:16 crc kubenswrapper[4751]: I1002 11:38:16.457605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerDied","Data":"c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1"} Oct 02 11:38:18 crc kubenswrapper[4751]: I1002 11:38:18.490622 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerStarted","Data":"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0"} Oct 02 11:38:18 crc kubenswrapper[4751]: I1002 11:38:18.508851 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-89sqb" podStartSLOduration=3.6032417580000002 podStartE2EDuration="6.50883517s" podCreationTimestamp="2025-10-02 11:38:12 +0000 UTC" firstStartedPulling="2025-10-02 11:38:14.4353318 +0000 UTC m=+2776.489558250" lastFinishedPulling="2025-10-02 11:38:17.340925212 +0000 UTC m=+2779.395151662" observedRunningTime="2025-10-02 11:38:18.5066026 +0000 UTC m=+2780.560829080" watchObservedRunningTime="2025-10-02 11:38:18.50883517 +0000 UTC m=+2780.563061620" Oct 02 11:38:23 crc kubenswrapper[4751]: I1002 11:38:23.199539 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:23 crc kubenswrapper[4751]: I1002 11:38:23.200112 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:23 crc kubenswrapper[4751]: I1002 11:38:23.252344 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:23 crc kubenswrapper[4751]: I1002 11:38:23.593464 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:23 crc kubenswrapper[4751]: I1002 11:38:23.642184 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:25 crc kubenswrapper[4751]: I1002 11:38:25.544905 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-89sqb" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="registry-server" containerID="cri-o://013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0" gracePeriod=2 Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.025005 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.074250 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content\") pod \"92555bc2-e96e-42d7-825b-3b7dd66a883d\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.074364 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb62g\" (UniqueName: \"kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g\") pod \"92555bc2-e96e-42d7-825b-3b7dd66a883d\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.074447 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities\") pod \"92555bc2-e96e-42d7-825b-3b7dd66a883d\" (UID: \"92555bc2-e96e-42d7-825b-3b7dd66a883d\") " Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.075820 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities" (OuterVolumeSpecName: "utilities") pod "92555bc2-e96e-42d7-825b-3b7dd66a883d" (UID: "92555bc2-e96e-42d7-825b-3b7dd66a883d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.086580 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g" (OuterVolumeSpecName: "kube-api-access-rb62g") pod "92555bc2-e96e-42d7-825b-3b7dd66a883d" (UID: "92555bc2-e96e-42d7-825b-3b7dd66a883d"). InnerVolumeSpecName "kube-api-access-rb62g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.127438 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92555bc2-e96e-42d7-825b-3b7dd66a883d" (UID: "92555bc2-e96e-42d7-825b-3b7dd66a883d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.176367 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.176417 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb62g\" (UniqueName: \"kubernetes.io/projected/92555bc2-e96e-42d7-825b-3b7dd66a883d-kube-api-access-rb62g\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.176433 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92555bc2-e96e-42d7-825b-3b7dd66a883d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.560262 4751 generic.go:334] "Generic (PLEG): container finished" podID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerID="013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0" exitCode=0 Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.560315 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerDied","Data":"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0"} Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.560346 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-89sqb" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.560391 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-89sqb" event={"ID":"92555bc2-e96e-42d7-825b-3b7dd66a883d","Type":"ContainerDied","Data":"558e172ee757ebc6b1444bfaf93d16a67a1461ec97d9016150e2d613c7ae22cb"} Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.560461 4751 scope.go:117] "RemoveContainer" containerID="013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.595997 4751 scope.go:117] "RemoveContainer" containerID="c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.604123 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.623492 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-89sqb"] Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.646250 4751 scope.go:117] "RemoveContainer" containerID="164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.679566 4751 scope.go:117] "RemoveContainer" containerID="013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0" Oct 02 11:38:26 crc kubenswrapper[4751]: E1002 11:38:26.680620 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0\": container with ID starting with 013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0 not found: ID does not exist" containerID="013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.680703 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0"} err="failed to get container status \"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0\": rpc error: code = NotFound desc = could not find container \"013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0\": container with ID starting with 013295dcd80ba03db951414d72c4147b0c75d1ddb9b09bcf977af856d31caec0 not found: ID does not exist" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.680770 4751 scope.go:117] "RemoveContainer" containerID="c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1" Oct 02 11:38:26 crc kubenswrapper[4751]: E1002 11:38:26.681373 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1\": container with ID starting with c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1 not found: ID does not exist" containerID="c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.681416 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1"} err="failed to get container status \"c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1\": rpc error: code = NotFound desc = could not find container \"c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1\": container with ID starting with c5d8126fd59a63ff90ce1ddc8e83cf3d56b34c373e066fe816ddf9401317b2a1 not found: ID does not exist" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.681446 4751 scope.go:117] "RemoveContainer" containerID="164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6" Oct 02 11:38:26 crc kubenswrapper[4751]: E1002 11:38:26.681807 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6\": container with ID starting with 164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6 not found: ID does not exist" containerID="164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6" Oct 02 11:38:26 crc kubenswrapper[4751]: I1002 11:38:26.681838 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6"} err="failed to get container status \"164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6\": rpc error: code = NotFound desc = could not find container \"164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6\": container with ID starting with 164e50ff1a710c7dab00c8042bda7f68ec79a060d8da89e8c8a94b0c50c871b6 not found: ID does not exist" Oct 02 11:38:27 crc kubenswrapper[4751]: I1002 11:38:27.560066 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" path="/var/lib/kubelet/pods/92555bc2-e96e-42d7-825b-3b7dd66a883d/volumes" Oct 02 11:38:31 crc kubenswrapper[4751]: I1002 11:38:31.506996 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:38:31 crc kubenswrapper[4751]: I1002 11:38:31.507616 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:38:31 crc kubenswrapper[4751]: I1002 11:38:31.507659 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:38:31 crc kubenswrapper[4751]: I1002 11:38:31.508404 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:38:31 crc kubenswrapper[4751]: I1002 11:38:31.508469 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676" gracePeriod=600 Oct 02 11:38:32 crc kubenswrapper[4751]: I1002 11:38:32.620104 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676" exitCode=0 Oct 02 11:38:32 crc kubenswrapper[4751]: I1002 11:38:32.620164 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676"} Oct 02 11:38:32 crc kubenswrapper[4751]: I1002 11:38:32.620730 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc"} Oct 02 11:38:32 crc kubenswrapper[4751]: I1002 11:38:32.620749 4751 scope.go:117] "RemoveContainer" containerID="6715d300fd8136f129784c343433c35fb0df6c40586797db9e80290bfe143d8f" Oct 02 11:39:08 crc kubenswrapper[4751]: I1002 11:39:08.958183 4751 generic.go:334] "Generic (PLEG): container finished" podID="ef947945-9630-4475-84ea-0798e9707829" containerID="f641763d027dab4e2dc7bf127b27e279b5fadc2bb845159a7d308b65635c65f8" exitCode=0 Oct 02 11:39:08 crc kubenswrapper[4751]: I1002 11:39:08.958215 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" event={"ID":"ef947945-9630-4475-84ea-0798e9707829","Type":"ContainerDied","Data":"f641763d027dab4e2dc7bf127b27e279b5fadc2bb845159a7d308b65635c65f8"} Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.505270 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:09 crc kubenswrapper[4751]: E1002 11:39:09.505936 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="registry-server" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.505960 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="registry-server" Oct 02 11:39:09 crc kubenswrapper[4751]: E1002 11:39:09.505973 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="extract-content" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.505981 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="extract-content" Oct 02 11:39:09 crc kubenswrapper[4751]: E1002 11:39:09.505996 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="extract-utilities" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.506005 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="extract-utilities" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.506347 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="92555bc2-e96e-42d7-825b-3b7dd66a883d" containerName="registry-server" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.509336 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.514211 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.589836 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88lpp\" (UniqueName: \"kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.590289 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.590798 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.692487 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88lpp\" (UniqueName: \"kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.692678 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.692775 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.693207 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.693464 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.714374 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88lpp\" (UniqueName: \"kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp\") pod \"community-operators-84jnr\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:09 crc kubenswrapper[4751]: I1002 11:39:09.829668 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.360373 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:10 crc kubenswrapper[4751]: W1002 11:39:10.361060 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda41f2270_dab2_471f_a18a_9d316e592d4e.slice/crio-34fdfc4844d2dc47bb1052c3809e7b9fc1ffd016daf056c96b28d2715ce08dff WatchSource:0}: Error finding container 34fdfc4844d2dc47bb1052c3809e7b9fc1ffd016daf056c96b28d2715ce08dff: Status 404 returned error can't find the container with id 34fdfc4844d2dc47bb1052c3809e7b9fc1ffd016daf056c96b28d2715ce08dff Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.412213 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.610439 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key\") pod \"ef947945-9630-4475-84ea-0798e9707829\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.610910 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0\") pod \"ef947945-9630-4475-84ea-0798e9707829\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.610953 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory\") pod \"ef947945-9630-4475-84ea-0798e9707829\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.611058 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6cpj\" (UniqueName: \"kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj\") pod \"ef947945-9630-4475-84ea-0798e9707829\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.611259 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle\") pod \"ef947945-9630-4475-84ea-0798e9707829\" (UID: \"ef947945-9630-4475-84ea-0798e9707829\") " Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.616557 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj" (OuterVolumeSpecName: "kube-api-access-l6cpj") pod "ef947945-9630-4475-84ea-0798e9707829" (UID: "ef947945-9630-4475-84ea-0798e9707829"). InnerVolumeSpecName "kube-api-access-l6cpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.617412 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ef947945-9630-4475-84ea-0798e9707829" (UID: "ef947945-9630-4475-84ea-0798e9707829"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.644407 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory" (OuterVolumeSpecName: "inventory") pod "ef947945-9630-4475-84ea-0798e9707829" (UID: "ef947945-9630-4475-84ea-0798e9707829"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.647240 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ef947945-9630-4475-84ea-0798e9707829" (UID: "ef947945-9630-4475-84ea-0798e9707829"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.663106 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef947945-9630-4475-84ea-0798e9707829" (UID: "ef947945-9630-4475-84ea-0798e9707829"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.715360 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.715658 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6cpj\" (UniqueName: \"kubernetes.io/projected/ef947945-9630-4475-84ea-0798e9707829-kube-api-access-l6cpj\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.715730 4751 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.715786 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.716102 4751 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ef947945-9630-4475-84ea-0798e9707829-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.982467 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" event={"ID":"ef947945-9630-4475-84ea-0798e9707829","Type":"ContainerDied","Data":"2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54"} Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.982735 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f36a38832db2eb17a62502dd2919561b45df5c2d5cbcf152c54d059a9c0dd54" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.982507 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9" Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.984973 4751 generic.go:334] "Generic (PLEG): container finished" podID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerID="2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f" exitCode=0 Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.985007 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerDied","Data":"2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f"} Oct 02 11:39:10 crc kubenswrapper[4751]: I1002 11:39:10.985027 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerStarted","Data":"34fdfc4844d2dc47bb1052c3809e7b9fc1ffd016daf056c96b28d2715ce08dff"} Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.106692 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94"] Oct 02 11:39:11 crc kubenswrapper[4751]: E1002 11:39:11.107397 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef947945-9630-4475-84ea-0798e9707829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.107479 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef947945-9630-4475-84ea-0798e9707829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.107793 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef947945-9630-4475-84ea-0798e9707829" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.108821 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.111689 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.111984 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.112309 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.112368 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.112327 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.113946 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.115636 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94"] Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.116531 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227106 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227529 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227561 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227613 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227663 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrxbm\" (UniqueName: \"kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227713 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227773 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227809 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.227917 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330105 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330282 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330407 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330440 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330458 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330518 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrxbm\" (UniqueName: \"kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330536 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.330571 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.331418 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.335466 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.335825 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.335934 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.336056 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.336611 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.336970 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.338094 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.348838 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrxbm\" (UniqueName: \"kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnp94\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.435760 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.705286 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.707926 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.714751 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.839615 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk4dg\" (UniqueName: \"kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.839790 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.840208 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.905371 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.907424 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.941650 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.941900 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk4dg\" (UniqueName: \"kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.942031 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.942136 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.942549 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.963708 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk4dg\" (UniqueName: \"kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg\") pod \"redhat-operators-lgpbp\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.979322 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:11 crc kubenswrapper[4751]: I1002 11:39:11.995385 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94"] Oct 02 11:39:12 crc kubenswrapper[4751]: W1002 11:39:12.007412 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod745cbea8_a96d_45d6_baf9_13c0ef35df34.slice/crio-6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c WatchSource:0}: Error finding container 6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c: Status 404 returned error can't find the container with id 6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.032795 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.043684 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.043778 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.043827 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfjkl\" (UniqueName: \"kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.147041 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.147567 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.147623 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfjkl\" (UniqueName: \"kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.147802 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.148082 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.179390 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfjkl\" (UniqueName: \"kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl\") pod \"redhat-marketplace-lnzts\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.226799 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:12 crc kubenswrapper[4751]: W1002 11:39:12.545735 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cf2e86e_b3d6_4a88_b01c_5355e27f01d7.slice/crio-75ee32050c13928c4287fa3a86fe1164d94a6083a6eebad9e33a315cb5046c9e WatchSource:0}: Error finding container 75ee32050c13928c4287fa3a86fe1164d94a6083a6eebad9e33a315cb5046c9e: Status 404 returned error can't find the container with id 75ee32050c13928c4287fa3a86fe1164d94a6083a6eebad9e33a315cb5046c9e Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.550073 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:12 crc kubenswrapper[4751]: W1002 11:39:12.559255 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cd471f5_0bc0_42af_be29_cba905ad4e63.slice/crio-baa6e60091e5c5f9053f73d4cee08a521a91e6a3e0b3a42fa5b47db3063705b7 WatchSource:0}: Error finding container baa6e60091e5c5f9053f73d4cee08a521a91e6a3e0b3a42fa5b47db3063705b7: Status 404 returned error can't find the container with id baa6e60091e5c5f9053f73d4cee08a521a91e6a3e0b3a42fa5b47db3063705b7 Oct 02 11:39:12 crc kubenswrapper[4751]: I1002 11:39:12.568399 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.017078 4751 generic.go:334] "Generic (PLEG): container finished" podID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerID="da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f" exitCode=0 Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.017144 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerDied","Data":"da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.024533 4751 generic.go:334] "Generic (PLEG): container finished" podID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerID="e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd" exitCode=0 Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.024605 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerDied","Data":"e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.024640 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerStarted","Data":"75ee32050c13928c4287fa3a86fe1164d94a6083a6eebad9e33a315cb5046c9e"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.034491 4751 generic.go:334] "Generic (PLEG): container finished" podID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerID="d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90" exitCode=0 Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.034613 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerDied","Data":"d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.034646 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerStarted","Data":"baa6e60091e5c5f9053f73d4cee08a521a91e6a3e0b3a42fa5b47db3063705b7"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.047712 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" event={"ID":"745cbea8-a96d-45d6-baf9-13c0ef35df34","Type":"ContainerStarted","Data":"5e618fca8fe84f245cb7c7f63c36b2bee0c330258b95d119d07977f9c1ec3062"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.047763 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" event={"ID":"745cbea8-a96d-45d6-baf9-13c0ef35df34","Type":"ContainerStarted","Data":"6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c"} Oct 02 11:39:13 crc kubenswrapper[4751]: I1002 11:39:13.098222 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" podStartSLOduration=1.76168715 podStartE2EDuration="2.098201717s" podCreationTimestamp="2025-10-02 11:39:11 +0000 UTC" firstStartedPulling="2025-10-02 11:39:12.013123266 +0000 UTC m=+2834.067349716" lastFinishedPulling="2025-10-02 11:39:12.349637833 +0000 UTC m=+2834.403864283" observedRunningTime="2025-10-02 11:39:13.09160215 +0000 UTC m=+2835.145828620" watchObservedRunningTime="2025-10-02 11:39:13.098201717 +0000 UTC m=+2835.152428177" Oct 02 11:39:14 crc kubenswrapper[4751]: I1002 11:39:14.062247 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerStarted","Data":"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a"} Oct 02 11:39:14 crc kubenswrapper[4751]: I1002 11:39:14.066545 4751 generic.go:334] "Generic (PLEG): container finished" podID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerID="b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656" exitCode=0 Oct 02 11:39:14 crc kubenswrapper[4751]: I1002 11:39:14.066632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerDied","Data":"b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656"} Oct 02 11:39:14 crc kubenswrapper[4751]: I1002 11:39:14.088379 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-84jnr" podStartSLOduration=2.535747943 podStartE2EDuration="5.088360446s" podCreationTimestamp="2025-10-02 11:39:09 +0000 UTC" firstStartedPulling="2025-10-02 11:39:10.986649731 +0000 UTC m=+2833.040876181" lastFinishedPulling="2025-10-02 11:39:13.539262234 +0000 UTC m=+2835.593488684" observedRunningTime="2025-10-02 11:39:14.083756072 +0000 UTC m=+2836.137982522" watchObservedRunningTime="2025-10-02 11:39:14.088360446 +0000 UTC m=+2836.142586896" Oct 02 11:39:15 crc kubenswrapper[4751]: I1002 11:39:15.080283 4751 generic.go:334] "Generic (PLEG): container finished" podID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerID="745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21" exitCode=0 Oct 02 11:39:15 crc kubenswrapper[4751]: I1002 11:39:15.081200 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerDied","Data":"745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21"} Oct 02 11:39:16 crc kubenswrapper[4751]: I1002 11:39:16.092527 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerStarted","Data":"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196"} Oct 02 11:39:16 crc kubenswrapper[4751]: I1002 11:39:16.095566 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerStarted","Data":"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c"} Oct 02 11:39:16 crc kubenswrapper[4751]: I1002 11:39:16.119949 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgpbp" podStartSLOduration=2.532478592 podStartE2EDuration="5.119931071s" podCreationTimestamp="2025-10-02 11:39:11 +0000 UTC" firstStartedPulling="2025-10-02 11:39:13.030561329 +0000 UTC m=+2835.084787779" lastFinishedPulling="2025-10-02 11:39:15.618013808 +0000 UTC m=+2837.672240258" observedRunningTime="2025-10-02 11:39:16.115823521 +0000 UTC m=+2838.170049971" watchObservedRunningTime="2025-10-02 11:39:16.119931071 +0000 UTC m=+2838.174157521" Oct 02 11:39:16 crc kubenswrapper[4751]: I1002 11:39:16.140534 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnzts" podStartSLOduration=3.164141094 podStartE2EDuration="5.140518875s" podCreationTimestamp="2025-10-02 11:39:11 +0000 UTC" firstStartedPulling="2025-10-02 11:39:13.038525123 +0000 UTC m=+2835.092751573" lastFinishedPulling="2025-10-02 11:39:15.014902894 +0000 UTC m=+2837.069129354" observedRunningTime="2025-10-02 11:39:16.135684255 +0000 UTC m=+2838.189910755" watchObservedRunningTime="2025-10-02 11:39:16.140518875 +0000 UTC m=+2838.194745335" Oct 02 11:39:19 crc kubenswrapper[4751]: I1002 11:39:19.830734 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:19 crc kubenswrapper[4751]: I1002 11:39:19.831359 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:19 crc kubenswrapper[4751]: I1002 11:39:19.874294 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:20 crc kubenswrapper[4751]: I1002 11:39:20.175568 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:20 crc kubenswrapper[4751]: I1002 11:39:20.498515 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.033763 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.033803 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.083047 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.147823 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-84jnr" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="registry-server" containerID="cri-o://8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a" gracePeriod=2 Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.195499 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.227452 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.227606 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.276572 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.703721 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.851683 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content\") pod \"a41f2270-dab2-471f-a18a-9d316e592d4e\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.851803 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities\") pod \"a41f2270-dab2-471f-a18a-9d316e592d4e\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.852063 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88lpp\" (UniqueName: \"kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp\") pod \"a41f2270-dab2-471f-a18a-9d316e592d4e\" (UID: \"a41f2270-dab2-471f-a18a-9d316e592d4e\") " Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.852465 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities" (OuterVolumeSpecName: "utilities") pod "a41f2270-dab2-471f-a18a-9d316e592d4e" (UID: "a41f2270-dab2-471f-a18a-9d316e592d4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.852748 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.857565 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp" (OuterVolumeSpecName: "kube-api-access-88lpp") pod "a41f2270-dab2-471f-a18a-9d316e592d4e" (UID: "a41f2270-dab2-471f-a18a-9d316e592d4e"). InnerVolumeSpecName "kube-api-access-88lpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.899755 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a41f2270-dab2-471f-a18a-9d316e592d4e" (UID: "a41f2270-dab2-471f-a18a-9d316e592d4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.954102 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a41f2270-dab2-471f-a18a-9d316e592d4e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:22 crc kubenswrapper[4751]: I1002 11:39:22.954146 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88lpp\" (UniqueName: \"kubernetes.io/projected/a41f2270-dab2-471f-a18a-9d316e592d4e-kube-api-access-88lpp\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.158526 4751 generic.go:334] "Generic (PLEG): container finished" podID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerID="8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a" exitCode=0 Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.158619 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84jnr" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.158670 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerDied","Data":"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a"} Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.158730 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84jnr" event={"ID":"a41f2270-dab2-471f-a18a-9d316e592d4e","Type":"ContainerDied","Data":"34fdfc4844d2dc47bb1052c3809e7b9fc1ffd016daf056c96b28d2715ce08dff"} Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.158756 4751 scope.go:117] "RemoveContainer" containerID="8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.190467 4751 scope.go:117] "RemoveContainer" containerID="da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.193906 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.206140 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-84jnr"] Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.216615 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.222239 4751 scope.go:117] "RemoveContainer" containerID="2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.261331 4751 scope.go:117] "RemoveContainer" containerID="8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a" Oct 02 11:39:23 crc kubenswrapper[4751]: E1002 11:39:23.262014 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a\": container with ID starting with 8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a not found: ID does not exist" containerID="8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.262062 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a"} err="failed to get container status \"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a\": rpc error: code = NotFound desc = could not find container \"8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a\": container with ID starting with 8760416dab81c4c273a1d0fd578819b619305cbaf90e0b7d27821c511322cf7a not found: ID does not exist" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.262093 4751 scope.go:117] "RemoveContainer" containerID="da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f" Oct 02 11:39:23 crc kubenswrapper[4751]: E1002 11:39:23.262604 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f\": container with ID starting with da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f not found: ID does not exist" containerID="da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.262670 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f"} err="failed to get container status \"da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f\": rpc error: code = NotFound desc = could not find container \"da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f\": container with ID starting with da56d5270b965dd77c67c17be243833b3c9f01a5d61a00b8641ad1253adc007f not found: ID does not exist" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.262708 4751 scope.go:117] "RemoveContainer" containerID="2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f" Oct 02 11:39:23 crc kubenswrapper[4751]: E1002 11:39:23.263112 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f\": container with ID starting with 2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f not found: ID does not exist" containerID="2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.263145 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f"} err="failed to get container status \"2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f\": rpc error: code = NotFound desc = could not find container \"2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f\": container with ID starting with 2d9ed34e05d11e313aee23b42796f8c66d215aa07611627f942b5f9b6d0d271f not found: ID does not exist" Oct 02 11:39:23 crc kubenswrapper[4751]: I1002 11:39:23.561193 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" path="/var/lib/kubelet/pods/a41f2270-dab2-471f-a18a-9d316e592d4e/volumes" Oct 02 11:39:24 crc kubenswrapper[4751]: I1002 11:39:24.493963 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:24 crc kubenswrapper[4751]: I1002 11:39:24.494230 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgpbp" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="registry-server" containerID="cri-o://4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196" gracePeriod=2 Oct 02 11:39:24 crc kubenswrapper[4751]: I1002 11:39:24.954517 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.094574 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content\") pod \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.094623 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk4dg\" (UniqueName: \"kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg\") pod \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.094831 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities\") pod \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\" (UID: \"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7\") " Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.095900 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities" (OuterVolumeSpecName: "utilities") pod "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" (UID: "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.103010 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg" (OuterVolumeSpecName: "kube-api-access-qk4dg") pod "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" (UID: "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7"). InnerVolumeSpecName "kube-api-access-qk4dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.178842 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" (UID: "7cf2e86e-b3d6-4a88-b01c-5355e27f01d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.181764 4751 generic.go:334] "Generic (PLEG): container finished" podID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerID="4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196" exitCode=0 Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.181834 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgpbp" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.181857 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerDied","Data":"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196"} Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.181906 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgpbp" event={"ID":"7cf2e86e-b3d6-4a88-b01c-5355e27f01d7","Type":"ContainerDied","Data":"75ee32050c13928c4287fa3a86fe1164d94a6083a6eebad9e33a315cb5046c9e"} Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.181927 4751 scope.go:117] "RemoveContainer" containerID="4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.196809 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.196843 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk4dg\" (UniqueName: \"kubernetes.io/projected/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-kube-api-access-qk4dg\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.196858 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.204649 4751 scope.go:117] "RemoveContainer" containerID="745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.217812 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.225433 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgpbp"] Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.248310 4751 scope.go:117] "RemoveContainer" containerID="e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.277452 4751 scope.go:117] "RemoveContainer" containerID="4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196" Oct 02 11:39:25 crc kubenswrapper[4751]: E1002 11:39:25.277913 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196\": container with ID starting with 4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196 not found: ID does not exist" containerID="4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.277964 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196"} err="failed to get container status \"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196\": rpc error: code = NotFound desc = could not find container \"4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196\": container with ID starting with 4e7e2c93eb9e7baaf291b39a0ecba129ba524c1246559a3351d91aa8df640196 not found: ID does not exist" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.278022 4751 scope.go:117] "RemoveContainer" containerID="745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21" Oct 02 11:39:25 crc kubenswrapper[4751]: E1002 11:39:25.278342 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21\": container with ID starting with 745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21 not found: ID does not exist" containerID="745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.278374 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21"} err="failed to get container status \"745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21\": rpc error: code = NotFound desc = could not find container \"745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21\": container with ID starting with 745c0e86b81cac5003ffc4370a2c1bbba7a9d8288b3a6fb16b78954a2bf6bf21 not found: ID does not exist" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.278405 4751 scope.go:117] "RemoveContainer" containerID="e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd" Oct 02 11:39:25 crc kubenswrapper[4751]: E1002 11:39:25.278729 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd\": container with ID starting with e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd not found: ID does not exist" containerID="e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.278763 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd"} err="failed to get container status \"e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd\": rpc error: code = NotFound desc = could not find container \"e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd\": container with ID starting with e77b65050db3c877966a442d4b52db6c2468680d360f8b3f9b89890c07f474dd not found: ID does not exist" Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.492406 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:25 crc kubenswrapper[4751]: I1002 11:39:25.568749 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" path="/var/lib/kubelet/pods/7cf2e86e-b3d6-4a88-b01c-5355e27f01d7/volumes" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.193617 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnzts" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="registry-server" containerID="cri-o://5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c" gracePeriod=2 Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.637051 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.723609 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfjkl\" (UniqueName: \"kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl\") pod \"4cd471f5-0bc0-42af-be29-cba905ad4e63\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.723658 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content\") pod \"4cd471f5-0bc0-42af-be29-cba905ad4e63\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.723890 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities\") pod \"4cd471f5-0bc0-42af-be29-cba905ad4e63\" (UID: \"4cd471f5-0bc0-42af-be29-cba905ad4e63\") " Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.724614 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities" (OuterVolumeSpecName: "utilities") pod "4cd471f5-0bc0-42af-be29-cba905ad4e63" (UID: "4cd471f5-0bc0-42af-be29-cba905ad4e63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.732405 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl" (OuterVolumeSpecName: "kube-api-access-cfjkl") pod "4cd471f5-0bc0-42af-be29-cba905ad4e63" (UID: "4cd471f5-0bc0-42af-be29-cba905ad4e63"). InnerVolumeSpecName "kube-api-access-cfjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.737077 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cd471f5-0bc0-42af-be29-cba905ad4e63" (UID: "4cd471f5-0bc0-42af-be29-cba905ad4e63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.826954 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.826994 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfjkl\" (UniqueName: \"kubernetes.io/projected/4cd471f5-0bc0-42af-be29-cba905ad4e63-kube-api-access-cfjkl\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:26 crc kubenswrapper[4751]: I1002 11:39:26.827005 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd471f5-0bc0-42af-be29-cba905ad4e63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.205222 4751 generic.go:334] "Generic (PLEG): container finished" podID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerID="5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c" exitCode=0 Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.205266 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerDied","Data":"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c"} Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.205280 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnzts" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.205329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnzts" event={"ID":"4cd471f5-0bc0-42af-be29-cba905ad4e63","Type":"ContainerDied","Data":"baa6e60091e5c5f9053f73d4cee08a521a91e6a3e0b3a42fa5b47db3063705b7"} Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.205355 4751 scope.go:117] "RemoveContainer" containerID="5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.228095 4751 scope.go:117] "RemoveContainer" containerID="b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.240477 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.248045 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnzts"] Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.275088 4751 scope.go:117] "RemoveContainer" containerID="d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.302798 4751 scope.go:117] "RemoveContainer" containerID="5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c" Oct 02 11:39:27 crc kubenswrapper[4751]: E1002 11:39:27.303256 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c\": container with ID starting with 5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c not found: ID does not exist" containerID="5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.303285 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c"} err="failed to get container status \"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c\": rpc error: code = NotFound desc = could not find container \"5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c\": container with ID starting with 5f465be4412772ff42475872f2a3263cef09e21dc186ed4cbbeb8ed9ab9cab3c not found: ID does not exist" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.303306 4751 scope.go:117] "RemoveContainer" containerID="b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656" Oct 02 11:39:27 crc kubenswrapper[4751]: E1002 11:39:27.303525 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656\": container with ID starting with b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656 not found: ID does not exist" containerID="b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.303559 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656"} err="failed to get container status \"b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656\": rpc error: code = NotFound desc = could not find container \"b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656\": container with ID starting with b99a388ff0433f270adf23833b4fceea35ce7f12fa69ec5ef1a77f2006c0c656 not found: ID does not exist" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.303570 4751 scope.go:117] "RemoveContainer" containerID="d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90" Oct 02 11:39:27 crc kubenswrapper[4751]: E1002 11:39:27.303891 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90\": container with ID starting with d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90 not found: ID does not exist" containerID="d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.303911 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90"} err="failed to get container status \"d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90\": rpc error: code = NotFound desc = could not find container \"d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90\": container with ID starting with d7b28bcca20d0fc1f8a2671a676a54f34203d2f88053bb5c89bb5f5f16cedb90 not found: ID does not exist" Oct 02 11:39:27 crc kubenswrapper[4751]: I1002 11:39:27.560092 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" path="/var/lib/kubelet/pods/4cd471f5-0bc0-42af-be29-cba905ad4e63/volumes" Oct 02 11:40:31 crc kubenswrapper[4751]: I1002 11:40:31.507563 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:40:31 crc kubenswrapper[4751]: I1002 11:40:31.508280 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:41:01 crc kubenswrapper[4751]: I1002 11:41:01.506804 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:41:01 crc kubenswrapper[4751]: I1002 11:41:01.507482 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:41:31 crc kubenswrapper[4751]: I1002 11:41:31.507402 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:41:31 crc kubenswrapper[4751]: I1002 11:41:31.508729 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:41:31 crc kubenswrapper[4751]: I1002 11:41:31.508794 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:41:31 crc kubenswrapper[4751]: I1002 11:41:31.509574 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:41:31 crc kubenswrapper[4751]: I1002 11:41:31.509628 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" gracePeriod=600 Oct 02 11:41:32 crc kubenswrapper[4751]: E1002 11:41:32.166133 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:41:32 crc kubenswrapper[4751]: I1002 11:41:32.342806 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" exitCode=0 Oct 02 11:41:32 crc kubenswrapper[4751]: I1002 11:41:32.342876 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc"} Oct 02 11:41:32 crc kubenswrapper[4751]: I1002 11:41:32.342933 4751 scope.go:117] "RemoveContainer" containerID="f285ed95518398110ac3bb673dcc78d0db62923115728ce7b30469767af4b676" Oct 02 11:41:32 crc kubenswrapper[4751]: I1002 11:41:32.343655 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:41:32 crc kubenswrapper[4751]: E1002 11:41:32.343953 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:41:44 crc kubenswrapper[4751]: I1002 11:41:44.550998 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:41:44 crc kubenswrapper[4751]: E1002 11:41:44.552023 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:41:57 crc kubenswrapper[4751]: I1002 11:41:57.550988 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:41:57 crc kubenswrapper[4751]: E1002 11:41:57.552922 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:42:11 crc kubenswrapper[4751]: I1002 11:42:11.551064 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:42:11 crc kubenswrapper[4751]: E1002 11:42:11.551955 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:42:23 crc kubenswrapper[4751]: I1002 11:42:23.550335 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:42:23 crc kubenswrapper[4751]: E1002 11:42:23.551242 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:42:27 crc kubenswrapper[4751]: I1002 11:42:27.830367 4751 generic.go:334] "Generic (PLEG): container finished" podID="745cbea8-a96d-45d6-baf9-13c0ef35df34" containerID="5e618fca8fe84f245cb7c7f63c36b2bee0c330258b95d119d07977f9c1ec3062" exitCode=2 Oct 02 11:42:27 crc kubenswrapper[4751]: I1002 11:42:27.830450 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" event={"ID":"745cbea8-a96d-45d6-baf9-13c0ef35df34","Type":"ContainerDied","Data":"5e618fca8fe84f245cb7c7f63c36b2bee0c330258b95d119d07977f9c1ec3062"} Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.244928 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.364925 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.364967 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365102 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365134 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrxbm\" (UniqueName: \"kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365150 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365187 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365277 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365299 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.365337 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0\") pod \"745cbea8-a96d-45d6-baf9-13c0ef35df34\" (UID: \"745cbea8-a96d-45d6-baf9-13c0ef35df34\") " Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.370612 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm" (OuterVolumeSpecName: "kube-api-access-rrxbm") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "kube-api-access-rrxbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.374746 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.393338 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.394493 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.395775 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.395841 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.396672 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.400300 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.404053 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory" (OuterVolumeSpecName: "inventory") pod "745cbea8-a96d-45d6-baf9-13c0ef35df34" (UID: "745cbea8-a96d-45d6-baf9-13c0ef35df34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467119 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467160 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467188 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467200 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrxbm\" (UniqueName: \"kubernetes.io/projected/745cbea8-a96d-45d6-baf9-13c0ef35df34-kube-api-access-rrxbm\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467212 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467222 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467233 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467242 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.467253 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/745cbea8-a96d-45d6-baf9-13c0ef35df34-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.853087 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" event={"ID":"745cbea8-a96d-45d6-baf9-13c0ef35df34","Type":"ContainerDied","Data":"6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c"} Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.853145 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6079ec123b44a5ee56124fdf7914cae1a4debdd15f0df055fca8af200c8bb73c" Oct 02 11:42:29 crc kubenswrapper[4751]: I1002 11:42:29.853226 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnp94" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.031202 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z"] Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032062 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032075 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032091 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032097 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032111 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032116 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032127 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032134 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032142 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032148 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032162 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032187 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032197 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032204 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032220 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032226 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="extract-content" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032261 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="745cbea8-a96d-45d6-baf9-13c0ef35df34" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032269 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="745cbea8-a96d-45d6-baf9-13c0ef35df34" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.032286 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032293 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="extract-utilities" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032503 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="a41f2270-dab2-471f-a18a-9d316e592d4e" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032526 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf2e86e-b3d6-4a88-b01c-5355e27f01d7" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032547 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="745cbea8-a96d-45d6-baf9-13c0ef35df34" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.032564 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd471f5-0bc0-42af-be29-cba905ad4e63" containerName="registry-server" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.033656 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.036364 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.036549 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.037254 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.037570 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.039132 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.039487 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.045818 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.047752 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z"] Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102620 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102661 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102681 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102726 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102775 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2zf9\" (UniqueName: \"kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.102849 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.103105 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.103137 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204410 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204481 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204515 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204550 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204591 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.205364 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.204624 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.205510 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.205604 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.205761 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2zf9\" (UniqueName: \"kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.210672 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.210799 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.211089 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.211323 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.211434 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.211650 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.212160 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.222547 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2zf9\" (UniqueName: \"kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kxw7z\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.355219 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.549698 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:42:37 crc kubenswrapper[4751]: E1002 11:42:37.550237 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.859449 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z"] Oct 02 11:42:37 crc kubenswrapper[4751]: I1002 11:42:37.926986 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" event={"ID":"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7","Type":"ContainerStarted","Data":"aad35d49ed92bb3b514afe1495b1473153047ab7df20f4642ab62b5eb2913099"} Oct 02 11:42:38 crc kubenswrapper[4751]: I1002 11:42:38.948251 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" event={"ID":"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7","Type":"ContainerStarted","Data":"a82bbf1b5de3545f065bb88f022b64ef8ff0c1107eed303c257fe576b6f94742"} Oct 02 11:42:38 crc kubenswrapper[4751]: I1002 11:42:38.965894 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" podStartSLOduration=1.787263603 podStartE2EDuration="1.965873025s" podCreationTimestamp="2025-10-02 11:42:37 +0000 UTC" firstStartedPulling="2025-10-02 11:42:37.870974591 +0000 UTC m=+3039.925201041" lastFinishedPulling="2025-10-02 11:42:38.049584003 +0000 UTC m=+3040.103810463" observedRunningTime="2025-10-02 11:42:38.964349325 +0000 UTC m=+3041.018575785" watchObservedRunningTime="2025-10-02 11:42:38.965873025 +0000 UTC m=+3041.020099485" Oct 02 11:42:52 crc kubenswrapper[4751]: I1002 11:42:52.549994 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:42:52 crc kubenswrapper[4751]: E1002 11:42:52.550763 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:43:06 crc kubenswrapper[4751]: I1002 11:43:06.550607 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:43:06 crc kubenswrapper[4751]: E1002 11:43:06.551542 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:43:21 crc kubenswrapper[4751]: I1002 11:43:21.549782 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:43:21 crc kubenswrapper[4751]: E1002 11:43:21.550621 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:43:36 crc kubenswrapper[4751]: I1002 11:43:36.550515 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:43:36 crc kubenswrapper[4751]: E1002 11:43:36.551356 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:43:44 crc kubenswrapper[4751]: I1002 11:43:44.512403 4751 generic.go:334] "Generic (PLEG): container finished" podID="aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" containerID="a82bbf1b5de3545f065bb88f022b64ef8ff0c1107eed303c257fe576b6f94742" exitCode=2 Oct 02 11:43:44 crc kubenswrapper[4751]: I1002 11:43:44.512478 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" event={"ID":"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7","Type":"ContainerDied","Data":"a82bbf1b5de3545f065bb88f022b64ef8ff0c1107eed303c257fe576b6f94742"} Oct 02 11:43:45 crc kubenswrapper[4751]: I1002 11:43:45.900344 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.088199 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.088568 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.088671 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.088831 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2zf9\" (UniqueName: \"kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.088909 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.089017 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.089131 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.089239 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.089326 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle\") pod \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\" (UID: \"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7\") " Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.093554 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9" (OuterVolumeSpecName: "kube-api-access-h2zf9") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "kube-api-access-h2zf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.095601 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.121230 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.122197 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.122660 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.123349 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.127314 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory" (OuterVolumeSpecName: "inventory") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.127390 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.135710 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" (UID: "aeebd860-2b84-42c5-a4f4-2bd00d9a9de7"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191275 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191466 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191534 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191595 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2zf9\" (UniqueName: \"kubernetes.io/projected/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-kube-api-access-h2zf9\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191656 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191717 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191770 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191821 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.191879 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeebd860-2b84-42c5-a4f4-2bd00d9a9de7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.529782 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" event={"ID":"aeebd860-2b84-42c5-a4f4-2bd00d9a9de7","Type":"ContainerDied","Data":"aad35d49ed92bb3b514afe1495b1473153047ab7df20f4642ab62b5eb2913099"} Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.529827 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aad35d49ed92bb3b514afe1495b1473153047ab7df20f4642ab62b5eb2913099" Oct 02 11:43:46 crc kubenswrapper[4751]: I1002 11:43:46.530208 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kxw7z" Oct 02 11:43:51 crc kubenswrapper[4751]: I1002 11:43:51.552356 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:43:51 crc kubenswrapper[4751]: E1002 11:43:51.553479 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:44:02 crc kubenswrapper[4751]: I1002 11:44:02.550408 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:44:02 crc kubenswrapper[4751]: E1002 11:44:02.551854 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.028727 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v"] Oct 02 11:44:03 crc kubenswrapper[4751]: E1002 11:44:03.029233 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.029257 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.029507 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeebd860-2b84-42c5-a4f4-2bd00d9a9de7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.030346 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.033995 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.034064 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.034390 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.034538 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.034609 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.034690 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.037095 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.038064 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v"] Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196512 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196662 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196711 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196738 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196897 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.196950 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.197016 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.197059 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gcvm\" (UniqueName: \"kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.197221 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299378 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299438 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299469 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299523 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299552 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299591 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299642 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gcvm\" (UniqueName: \"kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299673 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.299727 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.300458 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.305663 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.306388 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.306694 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.307106 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.307209 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.307352 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.307599 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.317050 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gcvm\" (UniqueName: \"kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z2s8v\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.360284 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.847213 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v"] Oct 02 11:44:03 crc kubenswrapper[4751]: I1002 11:44:03.853124 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:44:04 crc kubenswrapper[4751]: I1002 11:44:04.682308 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" event={"ID":"ecb4700e-df83-4e6d-be73-132f8ca80b09","Type":"ContainerStarted","Data":"bc0fdbddac6228b9b1b1ad3aadaffff9c27f8137dc56248d83aa59ce23776c16"} Oct 02 11:44:04 crc kubenswrapper[4751]: I1002 11:44:04.682652 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" event={"ID":"ecb4700e-df83-4e6d-be73-132f8ca80b09","Type":"ContainerStarted","Data":"2bd7524dfd2e936d02e6786e35416fc7264c7eaddb476f52e1dbc55f83e532e3"} Oct 02 11:44:04 crc kubenswrapper[4751]: I1002 11:44:04.715984 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" podStartSLOduration=1.559796021 podStartE2EDuration="1.715959359s" podCreationTimestamp="2025-10-02 11:44:03 +0000 UTC" firstStartedPulling="2025-10-02 11:44:03.852820656 +0000 UTC m=+3125.907047116" lastFinishedPulling="2025-10-02 11:44:04.008984004 +0000 UTC m=+3126.063210454" observedRunningTime="2025-10-02 11:44:04.70407047 +0000 UTC m=+3126.758296940" watchObservedRunningTime="2025-10-02 11:44:04.715959359 +0000 UTC m=+3126.770185829" Oct 02 11:44:14 crc kubenswrapper[4751]: I1002 11:44:14.550087 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:44:14 crc kubenswrapper[4751]: E1002 11:44:14.550911 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:44:28 crc kubenswrapper[4751]: I1002 11:44:28.550762 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:44:28 crc kubenswrapper[4751]: E1002 11:44:28.551481 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:44:39 crc kubenswrapper[4751]: I1002 11:44:39.557826 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:44:39 crc kubenswrapper[4751]: E1002 11:44:39.558624 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:44:53 crc kubenswrapper[4751]: I1002 11:44:53.550994 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:44:53 crc kubenswrapper[4751]: E1002 11:44:53.551831 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.157089 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6"] Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.158873 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.163895 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.164875 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.173440 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6"] Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.333723 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.334068 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c2l4\" (UniqueName: \"kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.334273 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.437036 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.437080 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c2l4\" (UniqueName: \"kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.437149 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.438117 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.443055 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.453869 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c2l4\" (UniqueName: \"kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4\") pod \"collect-profiles-29323425-m4tc6\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.487701 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:00 crc kubenswrapper[4751]: I1002 11:45:00.917772 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6"] Oct 02 11:45:01 crc kubenswrapper[4751]: I1002 11:45:01.179054 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" event={"ID":"dea9dbc6-a752-4824-80d7-468fdb4c849d","Type":"ContainerStarted","Data":"73788cb9a99df4da9b1e54415bb556631b950d4a272ea0ab5acbfc2a632834f3"} Oct 02 11:45:01 crc kubenswrapper[4751]: I1002 11:45:01.179098 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" event={"ID":"dea9dbc6-a752-4824-80d7-468fdb4c849d","Type":"ContainerStarted","Data":"cdc436dfc89d79583c5f2776537249b5b8ab80e4ff7da63b32f53840526bbf3f"} Oct 02 11:45:01 crc kubenswrapper[4751]: I1002 11:45:01.204559 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" podStartSLOduration=1.204532228 podStartE2EDuration="1.204532228s" podCreationTimestamp="2025-10-02 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 11:45:01.197098798 +0000 UTC m=+3183.251325258" watchObservedRunningTime="2025-10-02 11:45:01.204532228 +0000 UTC m=+3183.258758688" Oct 02 11:45:02 crc kubenswrapper[4751]: I1002 11:45:02.188961 4751 generic.go:334] "Generic (PLEG): container finished" podID="dea9dbc6-a752-4824-80d7-468fdb4c849d" containerID="73788cb9a99df4da9b1e54415bb556631b950d4a272ea0ab5acbfc2a632834f3" exitCode=0 Oct 02 11:45:02 crc kubenswrapper[4751]: I1002 11:45:02.189028 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" event={"ID":"dea9dbc6-a752-4824-80d7-468fdb4c849d","Type":"ContainerDied","Data":"73788cb9a99df4da9b1e54415bb556631b950d4a272ea0ab5acbfc2a632834f3"} Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.530837 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.698740 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c2l4\" (UniqueName: \"kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4\") pod \"dea9dbc6-a752-4824-80d7-468fdb4c849d\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.698917 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume\") pod \"dea9dbc6-a752-4824-80d7-468fdb4c849d\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.699222 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume\") pod \"dea9dbc6-a752-4824-80d7-468fdb4c849d\" (UID: \"dea9dbc6-a752-4824-80d7-468fdb4c849d\") " Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.700126 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume" (OuterVolumeSpecName: "config-volume") pod "dea9dbc6-a752-4824-80d7-468fdb4c849d" (UID: "dea9dbc6-a752-4824-80d7-468fdb4c849d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.701525 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dea9dbc6-a752-4824-80d7-468fdb4c849d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.704570 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4" (OuterVolumeSpecName: "kube-api-access-2c2l4") pod "dea9dbc6-a752-4824-80d7-468fdb4c849d" (UID: "dea9dbc6-a752-4824-80d7-468fdb4c849d"). InnerVolumeSpecName "kube-api-access-2c2l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.705369 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dea9dbc6-a752-4824-80d7-468fdb4c849d" (UID: "dea9dbc6-a752-4824-80d7-468fdb4c849d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.803334 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c2l4\" (UniqueName: \"kubernetes.io/projected/dea9dbc6-a752-4824-80d7-468fdb4c849d-kube-api-access-2c2l4\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:03 crc kubenswrapper[4751]: I1002 11:45:03.803373 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dea9dbc6-a752-4824-80d7-468fdb4c849d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:04 crc kubenswrapper[4751]: I1002 11:45:04.206631 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" event={"ID":"dea9dbc6-a752-4824-80d7-468fdb4c849d","Type":"ContainerDied","Data":"cdc436dfc89d79583c5f2776537249b5b8ab80e4ff7da63b32f53840526bbf3f"} Oct 02 11:45:04 crc kubenswrapper[4751]: I1002 11:45:04.206673 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdc436dfc89d79583c5f2776537249b5b8ab80e4ff7da63b32f53840526bbf3f" Oct 02 11:45:04 crc kubenswrapper[4751]: I1002 11:45:04.206707 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323425-m4tc6" Oct 02 11:45:04 crc kubenswrapper[4751]: I1002 11:45:04.264408 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t"] Oct 02 11:45:04 crc kubenswrapper[4751]: I1002 11:45:04.271861 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-kxc4t"] Oct 02 11:45:05 crc kubenswrapper[4751]: I1002 11:45:05.561208 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c25d4150-62c8-4ec1-ae32-050a04eafd3e" path="/var/lib/kubelet/pods/c25d4150-62c8-4ec1-ae32-050a04eafd3e/volumes" Oct 02 11:45:06 crc kubenswrapper[4751]: I1002 11:45:06.550305 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:45:06 crc kubenswrapper[4751]: E1002 11:45:06.550865 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:45:10 crc kubenswrapper[4751]: I1002 11:45:10.280968 4751 generic.go:334] "Generic (PLEG): container finished" podID="ecb4700e-df83-4e6d-be73-132f8ca80b09" containerID="bc0fdbddac6228b9b1b1ad3aadaffff9c27f8137dc56248d83aa59ce23776c16" exitCode=2 Oct 02 11:45:10 crc kubenswrapper[4751]: I1002 11:45:10.281021 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" event={"ID":"ecb4700e-df83-4e6d-be73-132f8ca80b09","Type":"ContainerDied","Data":"bc0fdbddac6228b9b1b1ad3aadaffff9c27f8137dc56248d83aa59ce23776c16"} Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.675145 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.791808 4751 scope.go:117] "RemoveContainer" containerID="3dea4feb9fb491e1a03c581a7fa1f0ed4c011fb19ed0a232bcce1c2642878209" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845361 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845407 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845481 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845514 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845538 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845587 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845614 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845637 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gcvm\" (UniqueName: \"kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.845721 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key\") pod \"ecb4700e-df83-4e6d-be73-132f8ca80b09\" (UID: \"ecb4700e-df83-4e6d-be73-132f8ca80b09\") " Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.850891 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm" (OuterVolumeSpecName: "kube-api-access-6gcvm") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "kube-api-access-6gcvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.851362 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.870778 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.874297 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.875159 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory" (OuterVolumeSpecName: "inventory") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.875209 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.875633 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.876476 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.877231 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ecb4700e-df83-4e6d-be73-132f8ca80b09" (UID: "ecb4700e-df83-4e6d-be73-132f8ca80b09"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947388 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947417 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947427 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947438 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947446 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947455 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947463 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947471 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb4700e-df83-4e6d-be73-132f8ca80b09-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:11 crc kubenswrapper[4751]: I1002 11:45:11.947478 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gcvm\" (UniqueName: \"kubernetes.io/projected/ecb4700e-df83-4e6d-be73-132f8ca80b09-kube-api-access-6gcvm\") on node \"crc\" DevicePath \"\"" Oct 02 11:45:12 crc kubenswrapper[4751]: I1002 11:45:12.299516 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" event={"ID":"ecb4700e-df83-4e6d-be73-132f8ca80b09","Type":"ContainerDied","Data":"2bd7524dfd2e936d02e6786e35416fc7264c7eaddb476f52e1dbc55f83e532e3"} Oct 02 11:45:12 crc kubenswrapper[4751]: I1002 11:45:12.299770 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bd7524dfd2e936d02e6786e35416fc7264c7eaddb476f52e1dbc55f83e532e3" Oct 02 11:45:12 crc kubenswrapper[4751]: I1002 11:45:12.299589 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z2s8v" Oct 02 11:45:20 crc kubenswrapper[4751]: I1002 11:45:20.550343 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:45:20 crc kubenswrapper[4751]: E1002 11:45:20.551454 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:45:31 crc kubenswrapper[4751]: I1002 11:45:31.550468 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:45:31 crc kubenswrapper[4751]: E1002 11:45:31.551143 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:45:43 crc kubenswrapper[4751]: I1002 11:45:43.550348 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:45:43 crc kubenswrapper[4751]: E1002 11:45:43.551154 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.028060 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx"] Oct 02 11:45:49 crc kubenswrapper[4751]: E1002 11:45:49.028998 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea9dbc6-a752-4824-80d7-468fdb4c849d" containerName="collect-profiles" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.029010 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea9dbc6-a752-4824-80d7-468fdb4c849d" containerName="collect-profiles" Oct 02 11:45:49 crc kubenswrapper[4751]: E1002 11:45:49.029049 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb4700e-df83-4e6d-be73-132f8ca80b09" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.029055 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb4700e-df83-4e6d-be73-132f8ca80b09" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.029250 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb4700e-df83-4e6d-be73-132f8ca80b09" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.029335 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea9dbc6-a752-4824-80d7-468fdb4c849d" containerName="collect-profiles" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.030008 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.034227 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.035852 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.036708 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.036744 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.038013 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.038081 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.038582 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx"] Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.039970 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136273 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136337 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136429 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136453 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136494 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2lds\" (UniqueName: \"kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136525 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136580 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136610 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.136648 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238242 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238315 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238337 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238403 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238425 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238457 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2lds\" (UniqueName: \"kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238479 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238523 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.238543 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.242092 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246354 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246358 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246358 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246417 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246651 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.246719 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.249111 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.257444 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2lds\" (UniqueName: \"kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-pmzbx\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.350856 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:45:49 crc kubenswrapper[4751]: I1002 11:45:49.856700 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx"] Oct 02 11:45:50 crc kubenswrapper[4751]: I1002 11:45:50.652331 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" event={"ID":"f861cad6-acad-4877-8054-e19e17de22ad","Type":"ContainerStarted","Data":"55e2fc6ef7f4320ad56c59e03ea89d397052685e35db0479d847de0447f63a74"} Oct 02 11:45:50 crc kubenswrapper[4751]: I1002 11:45:50.652946 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" event={"ID":"f861cad6-acad-4877-8054-e19e17de22ad","Type":"ContainerStarted","Data":"f20527cb0dbae49dbd2b81d65514045b05e2a5a9e94ee74ada77c2bea1775a0a"} Oct 02 11:45:50 crc kubenswrapper[4751]: I1002 11:45:50.677686 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" podStartSLOduration=1.502465637 podStartE2EDuration="1.677665597s" podCreationTimestamp="2025-10-02 11:45:49 +0000 UTC" firstStartedPulling="2025-10-02 11:45:49.864044834 +0000 UTC m=+3231.918271284" lastFinishedPulling="2025-10-02 11:45:50.039244794 +0000 UTC m=+3232.093471244" observedRunningTime="2025-10-02 11:45:50.668087439 +0000 UTC m=+3232.722313909" watchObservedRunningTime="2025-10-02 11:45:50.677665597 +0000 UTC m=+3232.731892057" Oct 02 11:45:55 crc kubenswrapper[4751]: I1002 11:45:55.550630 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:45:55 crc kubenswrapper[4751]: E1002 11:45:55.551560 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:46:09 crc kubenswrapper[4751]: I1002 11:46:09.558945 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:46:09 crc kubenswrapper[4751]: E1002 11:46:09.559988 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:46:21 crc kubenswrapper[4751]: I1002 11:46:21.551647 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:46:21 crc kubenswrapper[4751]: E1002 11:46:21.552495 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:46:36 crc kubenswrapper[4751]: I1002 11:46:36.550876 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:46:37 crc kubenswrapper[4751]: I1002 11:46:37.080241 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb"} Oct 02 11:46:58 crc kubenswrapper[4751]: I1002 11:46:58.274551 4751 generic.go:334] "Generic (PLEG): container finished" podID="f861cad6-acad-4877-8054-e19e17de22ad" containerID="55e2fc6ef7f4320ad56c59e03ea89d397052685e35db0479d847de0447f63a74" exitCode=2 Oct 02 11:46:58 crc kubenswrapper[4751]: I1002 11:46:58.274619 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" event={"ID":"f861cad6-acad-4877-8054-e19e17de22ad","Type":"ContainerDied","Data":"55e2fc6ef7f4320ad56c59e03ea89d397052685e35db0479d847de0447f63a74"} Oct 02 11:46:59 crc kubenswrapper[4751]: E1002 11:46:59.492094 4751 info.go:109] Failed to get network devices: open /sys/class/net/f20527cb0dbae49/address: no such file or directory Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.680408 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863148 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863292 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863330 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2lds\" (UniqueName: \"kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863357 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863375 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863462 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863576 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863617 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.863681 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.872461 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds" (OuterVolumeSpecName: "kube-api-access-n2lds") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "kube-api-access-n2lds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.877667 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.893331 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.893940 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.894663 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory" (OuterVolumeSpecName: "inventory") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.899006 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.899218 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: E1002 11:46:59.899553 4751 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1 podName:f861cad6-acad-4877-8054-e19e17de22ad nodeName:}" failed. No retries permitted until 2025-10-02 11:47:00.399489342 +0000 UTC m=+3302.453715792 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "nova-cell1-compute-config-1" (UniqueName: "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad") : error deleting /var/lib/kubelet/pods/f861cad6-acad-4877-8054-e19e17de22ad/volume-subpaths: remove /var/lib/kubelet/pods/f861cad6-acad-4877-8054-e19e17de22ad/volume-subpaths: no such file or directory Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.902949 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965603 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965628 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f861cad6-acad-4877-8054-e19e17de22ad-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965639 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2lds\" (UniqueName: \"kubernetes.io/projected/f861cad6-acad-4877-8054-e19e17de22ad-kube-api-access-n2lds\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965648 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965657 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965664 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965673 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:46:59 crc kubenswrapper[4751]: I1002 11:46:59.965681 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.293025 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" event={"ID":"f861cad6-acad-4877-8054-e19e17de22ad","Type":"ContainerDied","Data":"f20527cb0dbae49dbd2b81d65514045b05e2a5a9e94ee74ada77c2bea1775a0a"} Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.293084 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-pmzbx" Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.293098 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20527cb0dbae49dbd2b81d65514045b05e2a5a9e94ee74ada77c2bea1775a0a" Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.474601 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") pod \"f861cad6-acad-4877-8054-e19e17de22ad\" (UID: \"f861cad6-acad-4877-8054-e19e17de22ad\") " Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.480790 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "f861cad6-acad-4877-8054-e19e17de22ad" (UID: "f861cad6-acad-4877-8054-e19e17de22ad"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:47:00 crc kubenswrapper[4751]: I1002 11:47:00.578244 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f861cad6-acad-4877-8054-e19e17de22ad-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.041239 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4"] Oct 02 11:48:18 crc kubenswrapper[4751]: E1002 11:48:18.042308 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f861cad6-acad-4877-8054-e19e17de22ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.042327 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="f861cad6-acad-4877-8054-e19e17de22ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.042569 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="f861cad6-acad-4877-8054-e19e17de22ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.043439 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.045828 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.046571 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.046771 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.047012 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.046606 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.047548 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.047954 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.054312 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4"] Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197098 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197225 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197269 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197376 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzbdm\" (UniqueName: \"kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197435 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197497 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197589 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197643 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.197750 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300127 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300303 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300400 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300454 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300546 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzbdm\" (UniqueName: \"kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300648 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300696 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300736 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.300772 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.302008 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.305873 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.306074 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.306646 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.307669 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.306818 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.309345 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.309464 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.317120 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzbdm\" (UniqueName: \"kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-24tk4\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.364318 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.865049 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4"] Oct 02 11:48:18 crc kubenswrapper[4751]: I1002 11:48:18.979211 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" event={"ID":"43932132-4f36-453c-82c7-8d48ced7e943","Type":"ContainerStarted","Data":"ff33eb7d22017a59f9436bbd60aa2ee82ead0dbea7321d984b65c86da33007ef"} Oct 02 11:48:19 crc kubenswrapper[4751]: I1002 11:48:19.989988 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" event={"ID":"43932132-4f36-453c-82c7-8d48ced7e943","Type":"ContainerStarted","Data":"88d28f3e5d90aa86f46d6d880a503e400e3f0700689069ccd4264e13bdf1c608"} Oct 02 11:48:20 crc kubenswrapper[4751]: I1002 11:48:20.019444 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" podStartSLOduration=1.8465052480000002 podStartE2EDuration="2.019416426s" podCreationTimestamp="2025-10-02 11:48:18 +0000 UTC" firstStartedPulling="2025-10-02 11:48:18.868439956 +0000 UTC m=+3380.922666406" lastFinishedPulling="2025-10-02 11:48:19.041351134 +0000 UTC m=+3381.095577584" observedRunningTime="2025-10-02 11:48:20.010532506 +0000 UTC m=+3382.064758966" watchObservedRunningTime="2025-10-02 11:48:20.019416426 +0000 UTC m=+3382.073642886" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.601680 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.604538 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.613773 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.744020 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.744097 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h986m\" (UniqueName: \"kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.744548 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.846452 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.846552 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.846602 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h986m\" (UniqueName: \"kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.846998 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.847228 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.870892 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h986m\" (UniqueName: \"kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m\") pod \"certified-operators-69sjf\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:57 crc kubenswrapper[4751]: I1002 11:48:57.940141 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:48:58 crc kubenswrapper[4751]: I1002 11:48:58.435254 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:48:59 crc kubenswrapper[4751]: I1002 11:48:59.333999 4751 generic.go:334] "Generic (PLEG): container finished" podID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerID="405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f" exitCode=0 Oct 02 11:48:59 crc kubenswrapper[4751]: I1002 11:48:59.334280 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerDied","Data":"405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f"} Oct 02 11:48:59 crc kubenswrapper[4751]: I1002 11:48:59.334322 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerStarted","Data":"bf9da4ea457b009d80554e48e1011110630376ee20309cd8ff8528e1ec864381"} Oct 02 11:49:00 crc kubenswrapper[4751]: I1002 11:49:00.345110 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerStarted","Data":"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9"} Oct 02 11:49:01 crc kubenswrapper[4751]: I1002 11:49:01.355498 4751 generic.go:334] "Generic (PLEG): container finished" podID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerID="b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9" exitCode=0 Oct 02 11:49:01 crc kubenswrapper[4751]: I1002 11:49:01.355549 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerDied","Data":"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9"} Oct 02 11:49:01 crc kubenswrapper[4751]: I1002 11:49:01.506969 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:49:01 crc kubenswrapper[4751]: I1002 11:49:01.507337 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:49:02 crc kubenswrapper[4751]: I1002 11:49:02.369031 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerStarted","Data":"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900"} Oct 02 11:49:02 crc kubenswrapper[4751]: I1002 11:49:02.395055 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-69sjf" podStartSLOduration=2.924713895 podStartE2EDuration="5.39503852s" podCreationTimestamp="2025-10-02 11:48:57 +0000 UTC" firstStartedPulling="2025-10-02 11:48:59.336073026 +0000 UTC m=+3421.390299476" lastFinishedPulling="2025-10-02 11:49:01.806397651 +0000 UTC m=+3423.860624101" observedRunningTime="2025-10-02 11:49:02.389412529 +0000 UTC m=+3424.443638999" watchObservedRunningTime="2025-10-02 11:49:02.39503852 +0000 UTC m=+3424.449264970" Oct 02 11:49:07 crc kubenswrapper[4751]: I1002 11:49:07.940402 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:07 crc kubenswrapper[4751]: I1002 11:49:07.941035 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:07 crc kubenswrapper[4751]: I1002 11:49:07.988204 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:08 crc kubenswrapper[4751]: I1002 11:49:08.504212 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:08 crc kubenswrapper[4751]: I1002 11:49:08.544878 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:49:10 crc kubenswrapper[4751]: I1002 11:49:10.438065 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-69sjf" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="registry-server" containerID="cri-o://20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900" gracePeriod=2 Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.080152 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.194567 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h986m\" (UniqueName: \"kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m\") pod \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.194680 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities\") pod \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.194769 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content\") pod \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\" (UID: \"ee12534c-d5c1-4eca-8c1c-df13d8c29689\") " Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.195782 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities" (OuterVolumeSpecName: "utilities") pod "ee12534c-d5c1-4eca-8c1c-df13d8c29689" (UID: "ee12534c-d5c1-4eca-8c1c-df13d8c29689"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.199362 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m" (OuterVolumeSpecName: "kube-api-access-h986m") pod "ee12534c-d5c1-4eca-8c1c-df13d8c29689" (UID: "ee12534c-d5c1-4eca-8c1c-df13d8c29689"). InnerVolumeSpecName "kube-api-access-h986m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.239353 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee12534c-d5c1-4eca-8c1c-df13d8c29689" (UID: "ee12534c-d5c1-4eca-8c1c-df13d8c29689"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.296974 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h986m\" (UniqueName: \"kubernetes.io/projected/ee12534c-d5c1-4eca-8c1c-df13d8c29689-kube-api-access-h986m\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.297247 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.297317 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee12534c-d5c1-4eca-8c1c-df13d8c29689-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.446819 4751 generic.go:334] "Generic (PLEG): container finished" podID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerID="20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900" exitCode=0 Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.446866 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerDied","Data":"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900"} Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.446888 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69sjf" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.446901 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69sjf" event={"ID":"ee12534c-d5c1-4eca-8c1c-df13d8c29689","Type":"ContainerDied","Data":"bf9da4ea457b009d80554e48e1011110630376ee20309cd8ff8528e1ec864381"} Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.446948 4751 scope.go:117] "RemoveContainer" containerID="20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.480337 4751 scope.go:117] "RemoveContainer" containerID="b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.484036 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.492404 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-69sjf"] Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.512035 4751 scope.go:117] "RemoveContainer" containerID="405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.549123 4751 scope.go:117] "RemoveContainer" containerID="20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900" Oct 02 11:49:11 crc kubenswrapper[4751]: E1002 11:49:11.550558 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900\": container with ID starting with 20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900 not found: ID does not exist" containerID="20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.550607 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900"} err="failed to get container status \"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900\": rpc error: code = NotFound desc = could not find container \"20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900\": container with ID starting with 20ecf0cc56befa6dcb103fe0e156c16fe82eba74c155c24490820e606953d900 not found: ID does not exist" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.550636 4751 scope.go:117] "RemoveContainer" containerID="b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9" Oct 02 11:49:11 crc kubenswrapper[4751]: E1002 11:49:11.551370 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9\": container with ID starting with b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9 not found: ID does not exist" containerID="b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.551408 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9"} err="failed to get container status \"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9\": rpc error: code = NotFound desc = could not find container \"b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9\": container with ID starting with b810f98c7c087699a8ab7a5fc2553b8b0ddc5f4c6a8090aab9fafb1cc7b8ece9 not found: ID does not exist" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.551428 4751 scope.go:117] "RemoveContainer" containerID="405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f" Oct 02 11:49:11 crc kubenswrapper[4751]: E1002 11:49:11.551727 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f\": container with ID starting with 405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f not found: ID does not exist" containerID="405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.551754 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f"} err="failed to get container status \"405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f\": rpc error: code = NotFound desc = could not find container \"405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f\": container with ID starting with 405de01bf9316c4d075d27fa76ebf5be3d8dbe2197ba30e1dc8b2a6ccc34076f not found: ID does not exist" Oct 02 11:49:11 crc kubenswrapper[4751]: I1002 11:49:11.563820 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" path="/var/lib/kubelet/pods/ee12534c-d5c1-4eca-8c1c-df13d8c29689/volumes" Oct 02 11:49:26 crc kubenswrapper[4751]: I1002 11:49:26.583696 4751 generic.go:334] "Generic (PLEG): container finished" podID="43932132-4f36-453c-82c7-8d48ced7e943" containerID="88d28f3e5d90aa86f46d6d880a503e400e3f0700689069ccd4264e13bdf1c608" exitCode=2 Oct 02 11:49:26 crc kubenswrapper[4751]: I1002 11:49:26.583774 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" event={"ID":"43932132-4f36-453c-82c7-8d48ced7e943","Type":"ContainerDied","Data":"88d28f3e5d90aa86f46d6d880a503e400e3f0700689069ccd4264e13bdf1c608"} Oct 02 11:49:27 crc kubenswrapper[4751]: I1002 11:49:27.975456 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.002950 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003057 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzbdm\" (UniqueName: \"kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003101 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003190 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003343 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003571 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003610 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003648 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.003760 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key\") pod \"43932132-4f36-453c-82c7-8d48ced7e943\" (UID: \"43932132-4f36-453c-82c7-8d48ced7e943\") " Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.026915 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm" (OuterVolumeSpecName: "kube-api-access-nzbdm") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "kube-api-access-nzbdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.027018 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.033192 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory" (OuterVolumeSpecName: "inventory") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.035771 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.039062 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.048214 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.050489 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.050850 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.059159 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43932132-4f36-453c-82c7-8d48ced7e943" (UID: "43932132-4f36-453c-82c7-8d48ced7e943"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105863 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105902 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105968 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105978 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzbdm\" (UniqueName: \"kubernetes.io/projected/43932132-4f36-453c-82c7-8d48ced7e943-kube-api-access-nzbdm\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105987 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43932132-4f36-453c-82c7-8d48ced7e943-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.105996 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.106005 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.106013 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.106022 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43932132-4f36-453c-82c7-8d48ced7e943-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.602528 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" event={"ID":"43932132-4f36-453c-82c7-8d48ced7e943","Type":"ContainerDied","Data":"ff33eb7d22017a59f9436bbd60aa2ee82ead0dbea7321d984b65c86da33007ef"} Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.602578 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-24tk4" Oct 02 11:49:28 crc kubenswrapper[4751]: I1002 11:49:28.602581 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff33eb7d22017a59f9436bbd60aa2ee82ead0dbea7321d984b65c86da33007ef" Oct 02 11:49:31 crc kubenswrapper[4751]: I1002 11:49:31.506780 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:49:31 crc kubenswrapper[4751]: I1002 11:49:31.507354 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.490137 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:49:54 crc kubenswrapper[4751]: E1002 11:49:54.491166 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43932132-4f36-453c-82c7-8d48ced7e943" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491204 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="43932132-4f36-453c-82c7-8d48ced7e943" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:49:54 crc kubenswrapper[4751]: E1002 11:49:54.491228 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="extract-utilities" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491235 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="extract-utilities" Oct 02 11:49:54 crc kubenswrapper[4751]: E1002 11:49:54.491245 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="extract-content" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491252 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="extract-content" Oct 02 11:49:54 crc kubenswrapper[4751]: E1002 11:49:54.491281 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="registry-server" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491289 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="registry-server" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491532 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee12534c-d5c1-4eca-8c1c-df13d8c29689" containerName="registry-server" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.491560 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="43932132-4f36-453c-82c7-8d48ced7e943" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.493234 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.499722 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.610027 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84fb8\" (UniqueName: \"kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.610120 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.610186 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.712554 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84fb8\" (UniqueName: \"kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.713013 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.713498 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.713561 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.713816 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.735090 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84fb8\" (UniqueName: \"kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8\") pod \"redhat-operators-hp2wt\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:54 crc kubenswrapper[4751]: I1002 11:49:54.812937 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:49:55 crc kubenswrapper[4751]: I1002 11:49:55.314692 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:49:55 crc kubenswrapper[4751]: I1002 11:49:55.855131 4751 generic.go:334] "Generic (PLEG): container finished" podID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerID="38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7" exitCode=0 Oct 02 11:49:55 crc kubenswrapper[4751]: I1002 11:49:55.855213 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerDied","Data":"38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7"} Oct 02 11:49:55 crc kubenswrapper[4751]: I1002 11:49:55.855503 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerStarted","Data":"a9e38a091a82d3f2a78d72e0d5fafb528ea52d2d6d269c09a97f2614b61b84d9"} Oct 02 11:49:55 crc kubenswrapper[4751]: I1002 11:49:55.857257 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.506835 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.507596 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.507664 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.508754 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.508845 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb" gracePeriod=600 Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.912066 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb" exitCode=0 Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.912129 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb"} Oct 02 11:50:01 crc kubenswrapper[4751]: I1002 11:50:01.912211 4751 scope.go:117] "RemoveContainer" containerID="ce1b666102a8db5f45eb8b2dbc6015ed1c73ff0a2ce0b56e3ad0eadc446c2dfc" Oct 02 11:50:02 crc kubenswrapper[4751]: I1002 11:50:02.923261 4751 generic.go:334] "Generic (PLEG): container finished" podID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerID="5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66" exitCode=0 Oct 02 11:50:02 crc kubenswrapper[4751]: I1002 11:50:02.923376 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerDied","Data":"5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66"} Oct 02 11:50:02 crc kubenswrapper[4751]: I1002 11:50:02.929511 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f"} Oct 02 11:50:03 crc kubenswrapper[4751]: I1002 11:50:03.942275 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerStarted","Data":"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff"} Oct 02 11:50:03 crc kubenswrapper[4751]: I1002 11:50:03.971552 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hp2wt" podStartSLOduration=2.48856662 podStartE2EDuration="9.971531597s" podCreationTimestamp="2025-10-02 11:49:54 +0000 UTC" firstStartedPulling="2025-10-02 11:49:55.857034425 +0000 UTC m=+3477.911260875" lastFinishedPulling="2025-10-02 11:50:03.339999402 +0000 UTC m=+3485.394225852" observedRunningTime="2025-10-02 11:50:03.956506482 +0000 UTC m=+3486.010732932" watchObservedRunningTime="2025-10-02 11:50:03.971531597 +0000 UTC m=+3486.025758047" Oct 02 11:50:04 crc kubenswrapper[4751]: I1002 11:50:04.813068 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:04 crc kubenswrapper[4751]: I1002 11:50:04.813112 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:05 crc kubenswrapper[4751]: I1002 11:50:05.856444 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hp2wt" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="registry-server" probeResult="failure" output=< Oct 02 11:50:05 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 11:50:05 crc kubenswrapper[4751]: > Oct 02 11:50:14 crc kubenswrapper[4751]: I1002 11:50:14.861605 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:14 crc kubenswrapper[4751]: I1002 11:50:14.915683 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:15 crc kubenswrapper[4751]: I1002 11:50:15.097117 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.057844 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hp2wt" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="registry-server" containerID="cri-o://01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff" gracePeriod=2 Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.497353 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.612421 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content\") pod \"eb587920-9a6f-47c9-bb89-d895a0b70d62\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.612478 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities\") pod \"eb587920-9a6f-47c9-bb89-d895a0b70d62\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.612588 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84fb8\" (UniqueName: \"kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8\") pod \"eb587920-9a6f-47c9-bb89-d895a0b70d62\" (UID: \"eb587920-9a6f-47c9-bb89-d895a0b70d62\") " Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.613586 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities" (OuterVolumeSpecName: "utilities") pod "eb587920-9a6f-47c9-bb89-d895a0b70d62" (UID: "eb587920-9a6f-47c9-bb89-d895a0b70d62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.617836 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8" (OuterVolumeSpecName: "kube-api-access-84fb8") pod "eb587920-9a6f-47c9-bb89-d895a0b70d62" (UID: "eb587920-9a6f-47c9-bb89-d895a0b70d62"). InnerVolumeSpecName "kube-api-access-84fb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.699627 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb587920-9a6f-47c9-bb89-d895a0b70d62" (UID: "eb587920-9a6f-47c9-bb89-d895a0b70d62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.714816 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.714855 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb587920-9a6f-47c9-bb89-d895a0b70d62-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:16 crc kubenswrapper[4751]: I1002 11:50:16.714869 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84fb8\" (UniqueName: \"kubernetes.io/projected/eb587920-9a6f-47c9-bb89-d895a0b70d62-kube-api-access-84fb8\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.069065 4751 generic.go:334] "Generic (PLEG): container finished" podID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerID="01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff" exitCode=0 Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.069137 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerDied","Data":"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff"} Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.069183 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp2wt" event={"ID":"eb587920-9a6f-47c9-bb89-d895a0b70d62","Type":"ContainerDied","Data":"a9e38a091a82d3f2a78d72e0d5fafb528ea52d2d6d269c09a97f2614b61b84d9"} Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.069206 4751 scope.go:117] "RemoveContainer" containerID="01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.069368 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp2wt" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.090795 4751 scope.go:117] "RemoveContainer" containerID="5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.104490 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.113063 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hp2wt"] Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.123561 4751 scope.go:117] "RemoveContainer" containerID="38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.153693 4751 scope.go:117] "RemoveContainer" containerID="01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff" Oct 02 11:50:17 crc kubenswrapper[4751]: E1002 11:50:17.154278 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff\": container with ID starting with 01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff not found: ID does not exist" containerID="01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.154324 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff"} err="failed to get container status \"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff\": rpc error: code = NotFound desc = could not find container \"01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff\": container with ID starting with 01f131d6fb5147423c9c9e03764c5e8dc98f9b3f78901d3a872aaad8ddf5b0ff not found: ID does not exist" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.154353 4751 scope.go:117] "RemoveContainer" containerID="5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66" Oct 02 11:50:17 crc kubenswrapper[4751]: E1002 11:50:17.154760 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66\": container with ID starting with 5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66 not found: ID does not exist" containerID="5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.154797 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66"} err="failed to get container status \"5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66\": rpc error: code = NotFound desc = could not find container \"5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66\": container with ID starting with 5a30623fe79a3384d8e3cb0fe58a8013d59a928649d9e9635a7f3000cf6dbf66 not found: ID does not exist" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.154822 4751 scope.go:117] "RemoveContainer" containerID="38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7" Oct 02 11:50:17 crc kubenswrapper[4751]: E1002 11:50:17.155152 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7\": container with ID starting with 38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7 not found: ID does not exist" containerID="38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.155209 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7"} err="failed to get container status \"38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7\": rpc error: code = NotFound desc = could not find container \"38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7\": container with ID starting with 38913bb11de9e48eb346192ad5dd4f819b4bb3ea777b0e7e561602dc6ff6ebe7 not found: ID does not exist" Oct 02 11:50:17 crc kubenswrapper[4751]: I1002 11:50:17.562926 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" path="/var/lib/kubelet/pods/eb587920-9a6f-47c9-bb89-d895a0b70d62/volumes" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.247927 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:27 crc kubenswrapper[4751]: E1002 11:50:27.249646 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="registry-server" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.249677 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="registry-server" Oct 02 11:50:27 crc kubenswrapper[4751]: E1002 11:50:27.249719 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="extract-utilities" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.249730 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="extract-utilities" Oct 02 11:50:27 crc kubenswrapper[4751]: E1002 11:50:27.249748 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="extract-content" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.249756 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="extract-content" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.250034 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb587920-9a6f-47c9-bb89-d895a0b70d62" containerName="registry-server" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.251913 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.265624 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.301029 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.301212 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j9t2\" (UniqueName: \"kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.301322 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.403291 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.403611 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j9t2\" (UniqueName: \"kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.403794 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.403831 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.404290 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.424114 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j9t2\" (UniqueName: \"kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2\") pod \"redhat-marketplace-86ps9\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:27 crc kubenswrapper[4751]: I1002 11:50:27.583293 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:28 crc kubenswrapper[4751]: I1002 11:50:28.011069 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:28 crc kubenswrapper[4751]: I1002 11:50:28.165153 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerStarted","Data":"ce49b1bd16ddc5c21c3a44bb79eef9dd3fb662cf7517334137449e8b1f429618"} Oct 02 11:50:29 crc kubenswrapper[4751]: I1002 11:50:29.175464 4751 generic.go:334] "Generic (PLEG): container finished" podID="27d05363-d0ce-4885-9714-8709a262a20a" containerID="085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac" exitCode=0 Oct 02 11:50:29 crc kubenswrapper[4751]: I1002 11:50:29.175527 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerDied","Data":"085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac"} Oct 02 11:50:31 crc kubenswrapper[4751]: I1002 11:50:31.198116 4751 generic.go:334] "Generic (PLEG): container finished" podID="27d05363-d0ce-4885-9714-8709a262a20a" containerID="e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72" exitCode=0 Oct 02 11:50:31 crc kubenswrapper[4751]: I1002 11:50:31.198215 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerDied","Data":"e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72"} Oct 02 11:50:32 crc kubenswrapper[4751]: I1002 11:50:32.211728 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerStarted","Data":"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7"} Oct 02 11:50:32 crc kubenswrapper[4751]: I1002 11:50:32.238126 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-86ps9" podStartSLOduration=2.765664413 podStartE2EDuration="5.238101285s" podCreationTimestamp="2025-10-02 11:50:27 +0000 UTC" firstStartedPulling="2025-10-02 11:50:29.177734663 +0000 UTC m=+3511.231961113" lastFinishedPulling="2025-10-02 11:50:31.650171535 +0000 UTC m=+3513.704397985" observedRunningTime="2025-10-02 11:50:32.229309348 +0000 UTC m=+3514.283535818" watchObservedRunningTime="2025-10-02 11:50:32.238101285 +0000 UTC m=+3514.292327735" Oct 02 11:50:37 crc kubenswrapper[4751]: I1002 11:50:37.584018 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:37 crc kubenswrapper[4751]: I1002 11:50:37.584591 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:37 crc kubenswrapper[4751]: I1002 11:50:37.629914 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:38 crc kubenswrapper[4751]: I1002 11:50:38.422308 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:38 crc kubenswrapper[4751]: I1002 11:50:38.468274 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.277019 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-86ps9" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="registry-server" containerID="cri-o://7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7" gracePeriod=2 Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.696124 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.837656 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j9t2\" (UniqueName: \"kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2\") pod \"27d05363-d0ce-4885-9714-8709a262a20a\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.837762 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content\") pod \"27d05363-d0ce-4885-9714-8709a262a20a\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.837804 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities\") pod \"27d05363-d0ce-4885-9714-8709a262a20a\" (UID: \"27d05363-d0ce-4885-9714-8709a262a20a\") " Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.838924 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities" (OuterVolumeSpecName: "utilities") pod "27d05363-d0ce-4885-9714-8709a262a20a" (UID: "27d05363-d0ce-4885-9714-8709a262a20a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.843836 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2" (OuterVolumeSpecName: "kube-api-access-4j9t2") pod "27d05363-d0ce-4885-9714-8709a262a20a" (UID: "27d05363-d0ce-4885-9714-8709a262a20a"). InnerVolumeSpecName "kube-api-access-4j9t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.852398 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27d05363-d0ce-4885-9714-8709a262a20a" (UID: "27d05363-d0ce-4885-9714-8709a262a20a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.939715 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j9t2\" (UniqueName: \"kubernetes.io/projected/27d05363-d0ce-4885-9714-8709a262a20a-kube-api-access-4j9t2\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.939772 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:40 crc kubenswrapper[4751]: I1002 11:50:40.939785 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d05363-d0ce-4885-9714-8709a262a20a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.290159 4751 generic.go:334] "Generic (PLEG): container finished" podID="27d05363-d0ce-4885-9714-8709a262a20a" containerID="7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7" exitCode=0 Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.290232 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86ps9" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.290250 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerDied","Data":"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7"} Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.291415 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86ps9" event={"ID":"27d05363-d0ce-4885-9714-8709a262a20a","Type":"ContainerDied","Data":"ce49b1bd16ddc5c21c3a44bb79eef9dd3fb662cf7517334137449e8b1f429618"} Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.291433 4751 scope.go:117] "RemoveContainer" containerID="7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.315025 4751 scope.go:117] "RemoveContainer" containerID="e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.326163 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.334870 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-86ps9"] Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.345674 4751 scope.go:117] "RemoveContainer" containerID="085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.386845 4751 scope.go:117] "RemoveContainer" containerID="7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7" Oct 02 11:50:41 crc kubenswrapper[4751]: E1002 11:50:41.387501 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7\": container with ID starting with 7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7 not found: ID does not exist" containerID="7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.387562 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7"} err="failed to get container status \"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7\": rpc error: code = NotFound desc = could not find container \"7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7\": container with ID starting with 7889c1b7e3247ff129398921859dcfd737b92a412e69085ba8eb696f3b0ff1f7 not found: ID does not exist" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.387603 4751 scope.go:117] "RemoveContainer" containerID="e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72" Oct 02 11:50:41 crc kubenswrapper[4751]: E1002 11:50:41.388133 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72\": container with ID starting with e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72 not found: ID does not exist" containerID="e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.388191 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72"} err="failed to get container status \"e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72\": rpc error: code = NotFound desc = could not find container \"e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72\": container with ID starting with e3fa519713495ee3b08016b2919c1f7da2f7b0262fd861a320143cc359112f72 not found: ID does not exist" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.388222 4751 scope.go:117] "RemoveContainer" containerID="085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac" Oct 02 11:50:41 crc kubenswrapper[4751]: E1002 11:50:41.388717 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac\": container with ID starting with 085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac not found: ID does not exist" containerID="085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.388761 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac"} err="failed to get container status \"085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac\": rpc error: code = NotFound desc = could not find container \"085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac\": container with ID starting with 085619f8a47fdc2f5610dfb5ea6a497f39dbb57fa0c323be006060e42cf8c7ac not found: ID does not exist" Oct 02 11:50:41 crc kubenswrapper[4751]: I1002 11:50:41.560476 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d05363-d0ce-4885-9714-8709a262a20a" path="/var/lib/kubelet/pods/27d05363-d0ce-4885-9714-8709a262a20a/volumes" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.032905 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd"] Oct 02 11:52:06 crc kubenswrapper[4751]: E1002 11:52:06.034787 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="extract-utilities" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.034830 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="extract-utilities" Oct 02 11:52:06 crc kubenswrapper[4751]: E1002 11:52:06.034856 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="registry-server" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.034867 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="registry-server" Oct 02 11:52:06 crc kubenswrapper[4751]: E1002 11:52:06.034883 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="extract-content" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.034896 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="extract-content" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.035258 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d05363-d0ce-4885-9714-8709a262a20a" containerName="registry-server" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.036395 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.039815 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.039910 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.039969 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.040040 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.040823 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.040889 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.042259 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.049416 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd"] Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165504 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165794 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165865 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165907 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165933 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k9h7\" (UniqueName: \"kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.165957 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.166005 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.166033 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.166050 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267370 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267458 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k9h7\" (UniqueName: \"kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267503 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267580 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267626 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267655 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267691 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267714 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.267807 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.268960 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.273859 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.274125 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.274513 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.274986 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.275192 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.276667 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.277635 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.284219 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k9h7\" (UniqueName: \"kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6hxbd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.366094 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:52:06 crc kubenswrapper[4751]: I1002 11:52:06.874637 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd"] Oct 02 11:52:07 crc kubenswrapper[4751]: I1002 11:52:07.045077 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" event={"ID":"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd","Type":"ContainerStarted","Data":"2ab293b445ac9520c84b55804cea277dba7cad4c6bb6758dac4e6f48463fbb17"} Oct 02 11:52:08 crc kubenswrapper[4751]: I1002 11:52:08.056141 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" event={"ID":"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd","Type":"ContainerStarted","Data":"f4127019bed70ba88f498536cae00e24fe49c8534d43b2892b4ea0ffdf423571"} Oct 02 11:52:08 crc kubenswrapper[4751]: I1002 11:52:08.091711 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" podStartSLOduration=1.9093210630000002 podStartE2EDuration="2.091675271s" podCreationTimestamp="2025-10-02 11:52:06 +0000 UTC" firstStartedPulling="2025-10-02 11:52:06.882549334 +0000 UTC m=+3608.936775774" lastFinishedPulling="2025-10-02 11:52:07.064903532 +0000 UTC m=+3609.119129982" observedRunningTime="2025-10-02 11:52:08.079599976 +0000 UTC m=+3610.133826436" watchObservedRunningTime="2025-10-02 11:52:08.091675271 +0000 UTC m=+3610.145901751" Oct 02 11:52:31 crc kubenswrapper[4751]: I1002 11:52:31.507111 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:52:31 crc kubenswrapper[4751]: I1002 11:52:31.507673 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:53:01 crc kubenswrapper[4751]: I1002 11:53:01.506907 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:53:01 crc kubenswrapper[4751]: I1002 11:53:01.507774 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:53:15 crc kubenswrapper[4751]: I1002 11:53:15.634124 4751 generic.go:334] "Generic (PLEG): container finished" podID="2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" containerID="f4127019bed70ba88f498536cae00e24fe49c8534d43b2892b4ea0ffdf423571" exitCode=2 Oct 02 11:53:15 crc kubenswrapper[4751]: I1002 11:53:15.634212 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" event={"ID":"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd","Type":"ContainerDied","Data":"f4127019bed70ba88f498536cae00e24fe49c8534d43b2892b4ea0ffdf423571"} Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.016779 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.209102 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.209151 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.209201 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.209321 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.209347 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.210069 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k9h7\" (UniqueName: \"kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.210319 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.210368 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.210453 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1\") pod \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\" (UID: \"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd\") " Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.215521 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7" (OuterVolumeSpecName: "kube-api-access-8k9h7") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "kube-api-access-8k9h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.215685 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.237905 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.238986 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.242371 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.244738 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.248191 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory" (OuterVolumeSpecName: "inventory") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.253317 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.255897 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" (UID: "2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.312927 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.312966 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.312980 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.312990 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.313000 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.313009 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k9h7\" (UniqueName: \"kubernetes.io/projected/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-kube-api-access-8k9h7\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.313017 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.313028 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.313039 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.652463 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" event={"ID":"2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd","Type":"ContainerDied","Data":"2ab293b445ac9520c84b55804cea277dba7cad4c6bb6758dac4e6f48463fbb17"} Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.652804 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ab293b445ac9520c84b55804cea277dba7cad4c6bb6758dac4e6f48463fbb17" Oct 02 11:53:17 crc kubenswrapper[4751]: I1002 11:53:17.652581 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6hxbd" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.507519 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.508445 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.508501 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.509337 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.509398 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" gracePeriod=600 Oct 02 11:53:31 crc kubenswrapper[4751]: E1002 11:53:31.628923 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.779486 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" exitCode=0 Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.779538 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f"} Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.779572 4751 scope.go:117] "RemoveContainer" containerID="5ba209fcee0e3832bcb002d7f32da2ee912114c991575f899a3c1c5e204dbbfb" Oct 02 11:53:31 crc kubenswrapper[4751]: I1002 11:53:31.780420 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:53:31 crc kubenswrapper[4751]: E1002 11:53:31.780740 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:53:42 crc kubenswrapper[4751]: I1002 11:53:42.550559 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:53:42 crc kubenswrapper[4751]: E1002 11:53:42.551413 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:53:54 crc kubenswrapper[4751]: I1002 11:53:54.550581 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:53:54 crc kubenswrapper[4751]: E1002 11:53:54.552309 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.184058 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:01 crc kubenswrapper[4751]: E1002 11:54:01.185079 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.185099 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.185404 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.187607 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.205930 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.288669 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.288711 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr5zp\" (UniqueName: \"kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.289618 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.391890 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.392009 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.392038 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr5zp\" (UniqueName: \"kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.392483 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.392560 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.415278 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr5zp\" (UniqueName: \"kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp\") pod \"community-operators-cjl8b\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:01 crc kubenswrapper[4751]: I1002 11:54:01.522554 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:02 crc kubenswrapper[4751]: I1002 11:54:02.017519 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:02 crc kubenswrapper[4751]: I1002 11:54:02.079976 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerStarted","Data":"b4bf8bca1e39970063917fab3755e8765b2a3d83be6abcb9d0da7704670d7bcf"} Oct 02 11:54:03 crc kubenswrapper[4751]: I1002 11:54:03.090047 4751 generic.go:334] "Generic (PLEG): container finished" podID="7572478e-1051-4332-88e1-94ce13d834b6" containerID="d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3" exitCode=0 Oct 02 11:54:03 crc kubenswrapper[4751]: I1002 11:54:03.090254 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerDied","Data":"d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3"} Oct 02 11:54:04 crc kubenswrapper[4751]: I1002 11:54:04.102621 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerStarted","Data":"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd"} Oct 02 11:54:05 crc kubenswrapper[4751]: I1002 11:54:05.113953 4751 generic.go:334] "Generic (PLEG): container finished" podID="7572478e-1051-4332-88e1-94ce13d834b6" containerID="77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd" exitCode=0 Oct 02 11:54:05 crc kubenswrapper[4751]: I1002 11:54:05.114029 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerDied","Data":"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd"} Oct 02 11:54:06 crc kubenswrapper[4751]: I1002 11:54:06.126355 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerStarted","Data":"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb"} Oct 02 11:54:06 crc kubenswrapper[4751]: I1002 11:54:06.148691 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cjl8b" podStartSLOduration=2.655251904 podStartE2EDuration="5.148667317s" podCreationTimestamp="2025-10-02 11:54:01 +0000 UTC" firstStartedPulling="2025-10-02 11:54:03.09221796 +0000 UTC m=+3725.146444410" lastFinishedPulling="2025-10-02 11:54:05.585633373 +0000 UTC m=+3727.639859823" observedRunningTime="2025-10-02 11:54:06.145556053 +0000 UTC m=+3728.199782513" watchObservedRunningTime="2025-10-02 11:54:06.148667317 +0000 UTC m=+3728.202893777" Oct 02 11:54:09 crc kubenswrapper[4751]: I1002 11:54:09.565520 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:54:09 crc kubenswrapper[4751]: E1002 11:54:09.566661 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:54:11 crc kubenswrapper[4751]: I1002 11:54:11.522718 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:11 crc kubenswrapper[4751]: I1002 11:54:11.523061 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:11 crc kubenswrapper[4751]: I1002 11:54:11.584479 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:12 crc kubenswrapper[4751]: I1002 11:54:12.238533 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:12 crc kubenswrapper[4751]: I1002 11:54:12.287142 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.202839 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cjl8b" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="registry-server" containerID="cri-o://8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb" gracePeriod=2 Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.634255 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.743896 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content\") pod \"7572478e-1051-4332-88e1-94ce13d834b6\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.744054 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr5zp\" (UniqueName: \"kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp\") pod \"7572478e-1051-4332-88e1-94ce13d834b6\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.744217 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities\") pod \"7572478e-1051-4332-88e1-94ce13d834b6\" (UID: \"7572478e-1051-4332-88e1-94ce13d834b6\") " Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.745921 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities" (OuterVolumeSpecName: "utilities") pod "7572478e-1051-4332-88e1-94ce13d834b6" (UID: "7572478e-1051-4332-88e1-94ce13d834b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.749571 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp" (OuterVolumeSpecName: "kube-api-access-xr5zp") pod "7572478e-1051-4332-88e1-94ce13d834b6" (UID: "7572478e-1051-4332-88e1-94ce13d834b6"). InnerVolumeSpecName "kube-api-access-xr5zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.802366 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7572478e-1051-4332-88e1-94ce13d834b6" (UID: "7572478e-1051-4332-88e1-94ce13d834b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.847272 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.847560 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7572478e-1051-4332-88e1-94ce13d834b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:14 crc kubenswrapper[4751]: I1002 11:54:14.847675 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr5zp\" (UniqueName: \"kubernetes.io/projected/7572478e-1051-4332-88e1-94ce13d834b6-kube-api-access-xr5zp\") on node \"crc\" DevicePath \"\"" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.219853 4751 generic.go:334] "Generic (PLEG): container finished" podID="7572478e-1051-4332-88e1-94ce13d834b6" containerID="8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb" exitCode=0 Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.219895 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerDied","Data":"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb"} Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.219910 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjl8b" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.219921 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjl8b" event={"ID":"7572478e-1051-4332-88e1-94ce13d834b6","Type":"ContainerDied","Data":"b4bf8bca1e39970063917fab3755e8765b2a3d83be6abcb9d0da7704670d7bcf"} Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.219937 4751 scope.go:117] "RemoveContainer" containerID="8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.247457 4751 scope.go:117] "RemoveContainer" containerID="77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.256678 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.265445 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cjl8b"] Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.284725 4751 scope.go:117] "RemoveContainer" containerID="d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.311534 4751 scope.go:117] "RemoveContainer" containerID="8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb" Oct 02 11:54:15 crc kubenswrapper[4751]: E1002 11:54:15.312007 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb\": container with ID starting with 8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb not found: ID does not exist" containerID="8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.312048 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb"} err="failed to get container status \"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb\": rpc error: code = NotFound desc = could not find container \"8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb\": container with ID starting with 8d0e26aaf1ed9f09aa59cdb73798c67abb2453e28af740448c3054560c5be9eb not found: ID does not exist" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.312075 4751 scope.go:117] "RemoveContainer" containerID="77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd" Oct 02 11:54:15 crc kubenswrapper[4751]: E1002 11:54:15.312522 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd\": container with ID starting with 77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd not found: ID does not exist" containerID="77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.312574 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd"} err="failed to get container status \"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd\": rpc error: code = NotFound desc = could not find container \"77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd\": container with ID starting with 77643de7942eb30a06ea59804361016165f538f12586259da507987857ccbfcd not found: ID does not exist" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.312603 4751 scope.go:117] "RemoveContainer" containerID="d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3" Oct 02 11:54:15 crc kubenswrapper[4751]: E1002 11:54:15.312848 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3\": container with ID starting with d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3 not found: ID does not exist" containerID="d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.312873 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3"} err="failed to get container status \"d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3\": rpc error: code = NotFound desc = could not find container \"d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3\": container with ID starting with d0d5bd8c8c1cd9371ce49fe890147b028ab0ec14ba55cb3e8137ce212005b7f3 not found: ID does not exist" Oct 02 11:54:15 crc kubenswrapper[4751]: I1002 11:54:15.564642 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7572478e-1051-4332-88e1-94ce13d834b6" path="/var/lib/kubelet/pods/7572478e-1051-4332-88e1-94ce13d834b6/volumes" Oct 02 11:54:21 crc kubenswrapper[4751]: I1002 11:54:21.550612 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:54:21 crc kubenswrapper[4751]: E1002 11:54:21.551424 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:54:35 crc kubenswrapper[4751]: I1002 11:54:35.555127 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:54:35 crc kubenswrapper[4751]: E1002 11:54:35.555929 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:54:46 crc kubenswrapper[4751]: I1002 11:54:46.550055 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:54:46 crc kubenswrapper[4751]: E1002 11:54:46.550844 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:54:58 crc kubenswrapper[4751]: I1002 11:54:58.550994 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:54:58 crc kubenswrapper[4751]: E1002 11:54:58.552289 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:55:11 crc kubenswrapper[4751]: I1002 11:55:11.550799 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:55:11 crc kubenswrapper[4751]: E1002 11:55:11.551751 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:55:25 crc kubenswrapper[4751]: I1002 11:55:25.550006 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:55:25 crc kubenswrapper[4751]: E1002 11:55:25.550767 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:55:40 crc kubenswrapper[4751]: I1002 11:55:40.550780 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:55:40 crc kubenswrapper[4751]: E1002 11:55:40.551987 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:55:53 crc kubenswrapper[4751]: I1002 11:55:53.551033 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:55:53 crc kubenswrapper[4751]: E1002 11:55:53.552642 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:56:04 crc kubenswrapper[4751]: I1002 11:56:04.549736 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:56:04 crc kubenswrapper[4751]: E1002 11:56:04.550629 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:56:18 crc kubenswrapper[4751]: I1002 11:56:18.550547 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:56:18 crc kubenswrapper[4751]: E1002 11:56:18.551240 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:56:31 crc kubenswrapper[4751]: I1002 11:56:31.550764 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:56:31 crc kubenswrapper[4751]: E1002 11:56:31.551645 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:56:42 crc kubenswrapper[4751]: I1002 11:56:42.550827 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:56:42 crc kubenswrapper[4751]: E1002 11:56:42.551715 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:56:57 crc kubenswrapper[4751]: I1002 11:56:57.550823 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:56:57 crc kubenswrapper[4751]: E1002 11:56:57.551633 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:57:08 crc kubenswrapper[4751]: I1002 11:57:08.550288 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:57:08 crc kubenswrapper[4751]: E1002 11:57:08.550904 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:57:21 crc kubenswrapper[4751]: I1002 11:57:21.551002 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:57:21 crc kubenswrapper[4751]: E1002 11:57:21.551903 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:57:34 crc kubenswrapper[4751]: I1002 11:57:34.550964 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:57:34 crc kubenswrapper[4751]: E1002 11:57:34.552199 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:57:45 crc kubenswrapper[4751]: I1002 11:57:45.550437 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:57:45 crc kubenswrapper[4751]: E1002 11:57:45.551278 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:58:00 crc kubenswrapper[4751]: I1002 11:58:00.550729 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:58:00 crc kubenswrapper[4751]: E1002 11:58:00.551519 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:58:15 crc kubenswrapper[4751]: I1002 11:58:15.550519 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:58:15 crc kubenswrapper[4751]: E1002 11:58:15.551529 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:58:28 crc kubenswrapper[4751]: I1002 11:58:28.550638 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:58:28 crc kubenswrapper[4751]: E1002 11:58:28.551537 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.064816 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg"] Oct 02 11:58:35 crc kubenswrapper[4751]: E1002 11:58:35.065948 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="extract-utilities" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.065967 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="extract-utilities" Oct 02 11:58:35 crc kubenswrapper[4751]: E1002 11:58:35.065986 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="extract-content" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.065995 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="extract-content" Oct 02 11:58:35 crc kubenswrapper[4751]: E1002 11:58:35.066019 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="registry-server" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.066028 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="registry-server" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.066354 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="7572478e-1051-4332-88e1-94ce13d834b6" containerName="registry-server" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.067406 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.068436 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg"] Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.070502 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.070609 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.070789 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.074243 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.074295 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lfrlk" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.074612 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.074610 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112143 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112401 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112652 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112694 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112727 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112748 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5xxb\" (UniqueName: \"kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112853 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112914 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.112948 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213775 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213865 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213888 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213915 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213937 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5xxb\" (UniqueName: \"kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.213983 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.214023 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.214055 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.214083 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.214853 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.219879 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.220162 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.220459 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.220592 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.220783 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.221190 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.221866 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.230614 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5xxb\" (UniqueName: \"kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qtgmg\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:35 crc kubenswrapper[4751]: I1002 11:58:35.391610 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:58:36 crc kubenswrapper[4751]: I1002 11:58:36.309446 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg"] Oct 02 11:58:36 crc kubenswrapper[4751]: I1002 11:58:36.311698 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:58:36 crc kubenswrapper[4751]: I1002 11:58:36.586638 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" event={"ID":"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac","Type":"ContainerStarted","Data":"1da5af845a07f0266e37f21cef79ecb78259298a8da81ac90db896310f883335"} Oct 02 11:58:37 crc kubenswrapper[4751]: I1002 11:58:37.599079 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" event={"ID":"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac","Type":"ContainerStarted","Data":"73907ee017870e7385a9b65d3cac4fb0d2e73e404df8dcf5abe7f1849e5402f1"} Oct 02 11:58:37 crc kubenswrapper[4751]: I1002 11:58:37.617753 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" podStartSLOduration=2.437471018 podStartE2EDuration="2.617730809s" podCreationTimestamp="2025-10-02 11:58:35 +0000 UTC" firstStartedPulling="2025-10-02 11:58:36.311490022 +0000 UTC m=+3998.365716472" lastFinishedPulling="2025-10-02 11:58:36.491749813 +0000 UTC m=+3998.545976263" observedRunningTime="2025-10-02 11:58:37.615071527 +0000 UTC m=+3999.669298017" watchObservedRunningTime="2025-10-02 11:58:37.617730809 +0000 UTC m=+3999.671957269" Oct 02 11:58:43 crc kubenswrapper[4751]: I1002 11:58:43.550688 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 11:58:44 crc kubenswrapper[4751]: I1002 11:58:44.670040 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e"} Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.028769 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.031099 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.042584 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.137863 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.137940 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlfx2\" (UniqueName: \"kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.138585 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.242255 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.242340 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlfx2\" (UniqueName: \"kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.242400 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.243202 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.243219 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.269679 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlfx2\" (UniqueName: \"kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2\") pod \"certified-operators-92jbd\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.350756 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:21 crc kubenswrapper[4751]: I1002 11:59:21.859971 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:22 crc kubenswrapper[4751]: I1002 11:59:22.017547 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerStarted","Data":"a1b6780a5298441431f1f170acba61d5719c24e453f4105d4cecfa3c0e151278"} Oct 02 11:59:23 crc kubenswrapper[4751]: I1002 11:59:23.027381 4751 generic.go:334] "Generic (PLEG): container finished" podID="954b0449-fc13-4522-9ddb-197dd7302663" containerID="4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3" exitCode=0 Oct 02 11:59:23 crc kubenswrapper[4751]: I1002 11:59:23.027432 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerDied","Data":"4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3"} Oct 02 11:59:25 crc kubenswrapper[4751]: I1002 11:59:25.048280 4751 generic.go:334] "Generic (PLEG): container finished" podID="954b0449-fc13-4522-9ddb-197dd7302663" containerID="98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318" exitCode=0 Oct 02 11:59:25 crc kubenswrapper[4751]: I1002 11:59:25.048469 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerDied","Data":"98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318"} Oct 02 11:59:26 crc kubenswrapper[4751]: I1002 11:59:26.059678 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerStarted","Data":"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754"} Oct 02 11:59:26 crc kubenswrapper[4751]: I1002 11:59:26.081571 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-92jbd" podStartSLOduration=2.654261836 podStartE2EDuration="5.081541405s" podCreationTimestamp="2025-10-02 11:59:21 +0000 UTC" firstStartedPulling="2025-10-02 11:59:23.029491897 +0000 UTC m=+4045.083718347" lastFinishedPulling="2025-10-02 11:59:25.456771466 +0000 UTC m=+4047.510997916" observedRunningTime="2025-10-02 11:59:26.074901176 +0000 UTC m=+4048.129127636" watchObservedRunningTime="2025-10-02 11:59:26.081541405 +0000 UTC m=+4048.135767855" Oct 02 11:59:31 crc kubenswrapper[4751]: I1002 11:59:31.351023 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:31 crc kubenswrapper[4751]: I1002 11:59:31.351685 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:31 crc kubenswrapper[4751]: I1002 11:59:31.412578 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:32 crc kubenswrapper[4751]: I1002 11:59:32.175935 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:32 crc kubenswrapper[4751]: I1002 11:59:32.226013 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.146888 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-92jbd" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="registry-server" containerID="cri-o://03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754" gracePeriod=2 Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.580651 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.716688 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content\") pod \"954b0449-fc13-4522-9ddb-197dd7302663\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.716815 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities\") pod \"954b0449-fc13-4522-9ddb-197dd7302663\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.716919 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlfx2\" (UniqueName: \"kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2\") pod \"954b0449-fc13-4522-9ddb-197dd7302663\" (UID: \"954b0449-fc13-4522-9ddb-197dd7302663\") " Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.717806 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities" (OuterVolumeSpecName: "utilities") pod "954b0449-fc13-4522-9ddb-197dd7302663" (UID: "954b0449-fc13-4522-9ddb-197dd7302663"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.718467 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.725437 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2" (OuterVolumeSpecName: "kube-api-access-mlfx2") pod "954b0449-fc13-4522-9ddb-197dd7302663" (UID: "954b0449-fc13-4522-9ddb-197dd7302663"). InnerVolumeSpecName "kube-api-access-mlfx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.767651 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "954b0449-fc13-4522-9ddb-197dd7302663" (UID: "954b0449-fc13-4522-9ddb-197dd7302663"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.820612 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlfx2\" (UniqueName: \"kubernetes.io/projected/954b0449-fc13-4522-9ddb-197dd7302663-kube-api-access-mlfx2\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:34 crc kubenswrapper[4751]: I1002 11:59:34.820646 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/954b0449-fc13-4522-9ddb-197dd7302663-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.156937 4751 generic.go:334] "Generic (PLEG): container finished" podID="954b0449-fc13-4522-9ddb-197dd7302663" containerID="03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754" exitCode=0 Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.156990 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerDied","Data":"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754"} Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.157024 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92jbd" event={"ID":"954b0449-fc13-4522-9ddb-197dd7302663","Type":"ContainerDied","Data":"a1b6780a5298441431f1f170acba61d5719c24e453f4105d4cecfa3c0e151278"} Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.157045 4751 scope.go:117] "RemoveContainer" containerID="03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.157051 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92jbd" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.175090 4751 scope.go:117] "RemoveContainer" containerID="98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.195055 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.203281 4751 scope.go:117] "RemoveContainer" containerID="4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.203743 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-92jbd"] Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.244330 4751 scope.go:117] "RemoveContainer" containerID="03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754" Oct 02 11:59:35 crc kubenswrapper[4751]: E1002 11:59:35.244743 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754\": container with ID starting with 03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754 not found: ID does not exist" containerID="03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.244780 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754"} err="failed to get container status \"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754\": rpc error: code = NotFound desc = could not find container \"03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754\": container with ID starting with 03c966ea13de4c219dbb8449c3d51800d0e076448c735cbf9d0727712568a754 not found: ID does not exist" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.244807 4751 scope.go:117] "RemoveContainer" containerID="98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318" Oct 02 11:59:35 crc kubenswrapper[4751]: E1002 11:59:35.245066 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318\": container with ID starting with 98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318 not found: ID does not exist" containerID="98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.245093 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318"} err="failed to get container status \"98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318\": rpc error: code = NotFound desc = could not find container \"98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318\": container with ID starting with 98944f902728b5c50a74a735f8bd5588f3c6ce7d79124109c509e3c60f698318 not found: ID does not exist" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.245110 4751 scope.go:117] "RemoveContainer" containerID="4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3" Oct 02 11:59:35 crc kubenswrapper[4751]: E1002 11:59:35.245568 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3\": container with ID starting with 4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3 not found: ID does not exist" containerID="4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.245627 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3"} err="failed to get container status \"4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3\": rpc error: code = NotFound desc = could not find container \"4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3\": container with ID starting with 4204f3421e0848ccf4d95d2b0722fc8882153a6babe924bbaff17552e1b712e3 not found: ID does not exist" Oct 02 11:59:35 crc kubenswrapper[4751]: I1002 11:59:35.568527 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954b0449-fc13-4522-9ddb-197dd7302663" path="/var/lib/kubelet/pods/954b0449-fc13-4522-9ddb-197dd7302663/volumes" Oct 02 11:59:50 crc kubenswrapper[4751]: I1002 11:59:50.297293 4751 generic.go:334] "Generic (PLEG): container finished" podID="839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" containerID="73907ee017870e7385a9b65d3cac4fb0d2e73e404df8dcf5abe7f1849e5402f1" exitCode=2 Oct 02 11:59:50 crc kubenswrapper[4751]: I1002 11:59:50.297378 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" event={"ID":"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac","Type":"ContainerDied","Data":"73907ee017870e7385a9b65d3cac4fb0d2e73e404df8dcf5abe7f1849e5402f1"} Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.716607 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860057 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860465 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860507 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5xxb\" (UniqueName: \"kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860534 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860573 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860652 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860704 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860770 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.860836 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0\") pod \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\" (UID: \"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac\") " Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.868446 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb" (OuterVolumeSpecName: "kube-api-access-r5xxb") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "kube-api-access-r5xxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.869423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.884014 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.885960 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.891023 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory" (OuterVolumeSpecName: "inventory") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.891628 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.906556 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.912945 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.915264 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" (UID: "839f86ef-2cd3-4c28-9b41-edc3b7dbcbac"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964061 4751 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964123 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964139 4751 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964154 4751 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964194 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5xxb\" (UniqueName: \"kubernetes.io/projected/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-kube-api-access-r5xxb\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964211 4751 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964225 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964238 4751 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:51 crc kubenswrapper[4751]: I1002 11:59:51.964250 4751 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/839f86ef-2cd3-4c28-9b41-edc3b7dbcbac-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 11:59:52 crc kubenswrapper[4751]: I1002 11:59:52.317814 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" event={"ID":"839f86ef-2cd3-4c28-9b41-edc3b7dbcbac","Type":"ContainerDied","Data":"1da5af845a07f0266e37f21cef79ecb78259298a8da81ac90db896310f883335"} Oct 02 11:59:52 crc kubenswrapper[4751]: I1002 11:59:52.317875 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da5af845a07f0266e37f21cef79ecb78259298a8da81ac90db896310f883335" Oct 02 11:59:52 crc kubenswrapper[4751]: I1002 11:59:52.317874 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qtgmg" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.142405 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh"] Oct 02 12:00:00 crc kubenswrapper[4751]: E1002 12:00:00.143338 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143354 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="extract-content" Oct 02 12:00:00 crc kubenswrapper[4751]: E1002 12:00:00.143374 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143381 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4751]: E1002 12:00:00.143418 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143426 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 12:00:00 crc kubenswrapper[4751]: E1002 12:00:00.143442 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143449 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="extract-utilities" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143672 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="954b0449-fc13-4522-9ddb-197dd7302663" containerName="registry-server" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.143684 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="839f86ef-2cd3-4c28-9b41-edc3b7dbcbac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.144343 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.147830 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.147916 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.156620 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh"] Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.230975 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwj8\" (UniqueName: \"kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.231107 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.231282 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.332933 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.333075 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.333252 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwj8\" (UniqueName: \"kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.333843 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.342433 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.355233 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwj8\" (UniqueName: \"kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8\") pod \"collect-profiles-29323440-vfprh\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.471517 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:00 crc kubenswrapper[4751]: I1002 12:00:00.922352 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh"] Oct 02 12:00:01 crc kubenswrapper[4751]: I1002 12:00:01.418499 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" event={"ID":"289d32f6-5de5-45a9-bfe3-96521e87adf6","Type":"ContainerStarted","Data":"21a422a27830ef50e0db627cc756867539125452a130429b29f80820527db091"} Oct 02 12:00:01 crc kubenswrapper[4751]: I1002 12:00:01.418578 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" event={"ID":"289d32f6-5de5-45a9-bfe3-96521e87adf6","Type":"ContainerStarted","Data":"1032f119fb53d05cc6c1ec9e0bafea534eeac4a173ec0557db98afd8b36dfc99"} Oct 02 12:00:01 crc kubenswrapper[4751]: I1002 12:00:01.445803 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" podStartSLOduration=1.445785553 podStartE2EDuration="1.445785553s" podCreationTimestamp="2025-10-02 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:00:01.438304931 +0000 UTC m=+4083.492531391" watchObservedRunningTime="2025-10-02 12:00:01.445785553 +0000 UTC m=+4083.500012003" Oct 02 12:00:02 crc kubenswrapper[4751]: I1002 12:00:02.438595 4751 generic.go:334] "Generic (PLEG): container finished" podID="289d32f6-5de5-45a9-bfe3-96521e87adf6" containerID="21a422a27830ef50e0db627cc756867539125452a130429b29f80820527db091" exitCode=0 Oct 02 12:00:02 crc kubenswrapper[4751]: I1002 12:00:02.438660 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" event={"ID":"289d32f6-5de5-45a9-bfe3-96521e87adf6","Type":"ContainerDied","Data":"21a422a27830ef50e0db627cc756867539125452a130429b29f80820527db091"} Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.774984 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.827823 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume\") pod \"289d32f6-5de5-45a9-bfe3-96521e87adf6\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.828102 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume\") pod \"289d32f6-5de5-45a9-bfe3-96521e87adf6\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.828375 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzwj8\" (UniqueName: \"kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8\") pod \"289d32f6-5de5-45a9-bfe3-96521e87adf6\" (UID: \"289d32f6-5de5-45a9-bfe3-96521e87adf6\") " Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.828634 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume" (OuterVolumeSpecName: "config-volume") pod "289d32f6-5de5-45a9-bfe3-96521e87adf6" (UID: "289d32f6-5de5-45a9-bfe3-96521e87adf6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.828988 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/289d32f6-5de5-45a9-bfe3-96521e87adf6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.835991 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8" (OuterVolumeSpecName: "kube-api-access-lzwj8") pod "289d32f6-5de5-45a9-bfe3-96521e87adf6" (UID: "289d32f6-5de5-45a9-bfe3-96521e87adf6"). InnerVolumeSpecName "kube-api-access-lzwj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.838333 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "289d32f6-5de5-45a9-bfe3-96521e87adf6" (UID: "289d32f6-5de5-45a9-bfe3-96521e87adf6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.930573 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzwj8\" (UniqueName: \"kubernetes.io/projected/289d32f6-5de5-45a9-bfe3-96521e87adf6-kube-api-access-lzwj8\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:03 crc kubenswrapper[4751]: I1002 12:00:03.930612 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/289d32f6-5de5-45a9-bfe3-96521e87adf6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:00:04 crc kubenswrapper[4751]: I1002 12:00:04.458643 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" event={"ID":"289d32f6-5de5-45a9-bfe3-96521e87adf6","Type":"ContainerDied","Data":"1032f119fb53d05cc6c1ec9e0bafea534eeac4a173ec0557db98afd8b36dfc99"} Oct 02 12:00:04 crc kubenswrapper[4751]: I1002 12:00:04.458699 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1032f119fb53d05cc6c1ec9e0bafea534eeac4a173ec0557db98afd8b36dfc99" Oct 02 12:00:04 crc kubenswrapper[4751]: I1002 12:00:04.458729 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323440-vfprh" Oct 02 12:00:04 crc kubenswrapper[4751]: I1002 12:00:04.517218 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p"] Oct 02 12:00:04 crc kubenswrapper[4751]: I1002 12:00:04.526189 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-r4q4p"] Oct 02 12:00:05 crc kubenswrapper[4751]: I1002 12:00:05.572878 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1bfc847-f75b-49bf-9249-b79d5a6139bd" path="/var/lib/kubelet/pods/b1bfc847-f75b-49bf-9249-b79d5a6139bd/volumes" Oct 02 12:00:12 crc kubenswrapper[4751]: I1002 12:00:12.149620 4751 scope.go:117] "RemoveContainer" containerID="3daa9e98d927dae2c70cb8883cedec1aa51cbe28e65b93038e46498c5080de24" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.152092 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323441-lp668"] Oct 02 12:01:00 crc kubenswrapper[4751]: E1002 12:01:00.154880 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="289d32f6-5de5-45a9-bfe3-96521e87adf6" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.155020 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="289d32f6-5de5-45a9-bfe3-96521e87adf6" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.155869 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="289d32f6-5de5-45a9-bfe3-96521e87adf6" containerName="collect-profiles" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.156854 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.176046 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323441-lp668"] Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.315845 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.316330 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.316528 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.316702 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4j5w\" (UniqueName: \"kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.418876 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.418999 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.419055 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.419093 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4j5w\" (UniqueName: \"kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.425226 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.425543 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.426561 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.435925 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4j5w\" (UniqueName: \"kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w\") pod \"keystone-cron-29323441-lp668\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.490527 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:00 crc kubenswrapper[4751]: I1002 12:01:00.927700 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323441-lp668"] Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.004045 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-lp668" event={"ID":"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96","Type":"ContainerStarted","Data":"7edf4b2b822be97196bfccbe5358de8b89fc684ed439049920919b4c653a69fd"} Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.185612 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.187665 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.199310 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.337099 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.337140 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.337679 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9xsv\" (UniqueName: \"kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.439481 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.439729 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.439850 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9xsv\" (UniqueName: \"kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.440582 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.440850 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.459854 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9xsv\" (UniqueName: \"kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv\") pod \"redhat-operators-x7j4k\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.508084 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.508193 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:01:01 crc kubenswrapper[4751]: I1002 12:01:01.516977 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:02 crc kubenswrapper[4751]: I1002 12:01:02.003246 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:02 crc kubenswrapper[4751]: I1002 12:01:02.021732 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-lp668" event={"ID":"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96","Type":"ContainerStarted","Data":"92ea7067747946af9bf6a5eb90df987bced3a672e919016f4ecda095b86e75f4"} Oct 02 12:01:02 crc kubenswrapper[4751]: I1002 12:01:02.057160 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323441-lp668" podStartSLOduration=2.057133564 podStartE2EDuration="2.057133564s" podCreationTimestamp="2025-10-02 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:01:02.046425785 +0000 UTC m=+4144.100652245" watchObservedRunningTime="2025-10-02 12:01:02.057133564 +0000 UTC m=+4144.111360024" Oct 02 12:01:03 crc kubenswrapper[4751]: I1002 12:01:03.032028 4751 generic.go:334] "Generic (PLEG): container finished" podID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerID="24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179" exitCode=0 Oct 02 12:01:03 crc kubenswrapper[4751]: I1002 12:01:03.034076 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerDied","Data":"24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179"} Oct 02 12:01:03 crc kubenswrapper[4751]: I1002 12:01:03.034403 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerStarted","Data":"212dabd87355741461f6a1b8cafdb080fe3e25e72ac0e1e0009c46007e4cedef"} Oct 02 12:01:04 crc kubenswrapper[4751]: I1002 12:01:04.044560 4751 generic.go:334] "Generic (PLEG): container finished" podID="eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" containerID="92ea7067747946af9bf6a5eb90df987bced3a672e919016f4ecda095b86e75f4" exitCode=0 Oct 02 12:01:04 crc kubenswrapper[4751]: I1002 12:01:04.044772 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-lp668" event={"ID":"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96","Type":"ContainerDied","Data":"92ea7067747946af9bf6a5eb90df987bced3a672e919016f4ecda095b86e75f4"} Oct 02 12:01:04 crc kubenswrapper[4751]: I1002 12:01:04.047858 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerStarted","Data":"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a"} Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.063103 4751 generic.go:334] "Generic (PLEG): container finished" podID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerID="3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a" exitCode=0 Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.063261 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerDied","Data":"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a"} Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.398693 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.537497 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle\") pod \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.538622 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys\") pod \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.538915 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4j5w\" (UniqueName: \"kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w\") pod \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.538939 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data\") pod \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\" (UID: \"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96\") " Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.545486 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w" (OuterVolumeSpecName: "kube-api-access-p4j5w") pod "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" (UID: "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96"). InnerVolumeSpecName "kube-api-access-p4j5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.546536 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" (UID: "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.568978 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" (UID: "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.630479 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data" (OuterVolumeSpecName: "config-data") pod "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" (UID: "eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.642069 4751 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.642115 4751 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.642130 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4j5w\" (UniqueName: \"kubernetes.io/projected/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-kube-api-access-p4j5w\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:05 crc kubenswrapper[4751]: I1002 12:01:05.642145 4751 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:06 crc kubenswrapper[4751]: I1002 12:01:06.075665 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerStarted","Data":"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6"} Oct 02 12:01:06 crc kubenswrapper[4751]: I1002 12:01:06.077862 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323441-lp668" event={"ID":"eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96","Type":"ContainerDied","Data":"7edf4b2b822be97196bfccbe5358de8b89fc684ed439049920919b4c653a69fd"} Oct 02 12:01:06 crc kubenswrapper[4751]: I1002 12:01:06.077939 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7edf4b2b822be97196bfccbe5358de8b89fc684ed439049920919b4c653a69fd" Oct 02 12:01:06 crc kubenswrapper[4751]: I1002 12:01:06.078065 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323441-lp668" Oct 02 12:01:06 crc kubenswrapper[4751]: I1002 12:01:06.121077 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7j4k" podStartSLOduration=2.580910188 podStartE2EDuration="5.121051301s" podCreationTimestamp="2025-10-02 12:01:01 +0000 UTC" firstStartedPulling="2025-10-02 12:01:03.035312224 +0000 UTC m=+4145.089538674" lastFinishedPulling="2025-10-02 12:01:05.575453317 +0000 UTC m=+4147.629679787" observedRunningTime="2025-10-02 12:01:06.097015893 +0000 UTC m=+4148.151242343" watchObservedRunningTime="2025-10-02 12:01:06.121051301 +0000 UTC m=+4148.175277751" Oct 02 12:01:11 crc kubenswrapper[4751]: I1002 12:01:11.518266 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:11 crc kubenswrapper[4751]: I1002 12:01:11.518954 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:11 crc kubenswrapper[4751]: I1002 12:01:11.713519 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:12 crc kubenswrapper[4751]: I1002 12:01:12.178958 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:12 crc kubenswrapper[4751]: I1002 12:01:12.220514 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.161075 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7j4k" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="registry-server" containerID="cri-o://521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6" gracePeriod=2 Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.611411 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.714972 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities\") pod \"dac93481-cb79-4f37-9da8-73cc941e9a63\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.715120 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content\") pod \"dac93481-cb79-4f37-9da8-73cc941e9a63\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.715793 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities" (OuterVolumeSpecName: "utilities") pod "dac93481-cb79-4f37-9da8-73cc941e9a63" (UID: "dac93481-cb79-4f37-9da8-73cc941e9a63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.716688 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9xsv\" (UniqueName: \"kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv\") pod \"dac93481-cb79-4f37-9da8-73cc941e9a63\" (UID: \"dac93481-cb79-4f37-9da8-73cc941e9a63\") " Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.717903 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.722963 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv" (OuterVolumeSpecName: "kube-api-access-m9xsv") pod "dac93481-cb79-4f37-9da8-73cc941e9a63" (UID: "dac93481-cb79-4f37-9da8-73cc941e9a63"). InnerVolumeSpecName "kube-api-access-m9xsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.791230 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dac93481-cb79-4f37-9da8-73cc941e9a63" (UID: "dac93481-cb79-4f37-9da8-73cc941e9a63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.819666 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dac93481-cb79-4f37-9da8-73cc941e9a63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:14 crc kubenswrapper[4751]: I1002 12:01:14.819703 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9xsv\" (UniqueName: \"kubernetes.io/projected/dac93481-cb79-4f37-9da8-73cc941e9a63-kube-api-access-m9xsv\") on node \"crc\" DevicePath \"\"" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.171363 4751 generic.go:334] "Generic (PLEG): container finished" podID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerID="521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6" exitCode=0 Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.171418 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerDied","Data":"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6"} Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.171463 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7j4k" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.171481 4751 scope.go:117] "RemoveContainer" containerID="521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.171468 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7j4k" event={"ID":"dac93481-cb79-4f37-9da8-73cc941e9a63","Type":"ContainerDied","Data":"212dabd87355741461f6a1b8cafdb080fe3e25e72ac0e1e0009c46007e4cedef"} Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.196399 4751 scope.go:117] "RemoveContainer" containerID="3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.216555 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.233228 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7j4k"] Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.234621 4751 scope.go:117] "RemoveContainer" containerID="24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.271543 4751 scope.go:117] "RemoveContainer" containerID="521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6" Oct 02 12:01:15 crc kubenswrapper[4751]: E1002 12:01:15.271975 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6\": container with ID starting with 521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6 not found: ID does not exist" containerID="521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.273124 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6"} err="failed to get container status \"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6\": rpc error: code = NotFound desc = could not find container \"521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6\": container with ID starting with 521c7e4e1256b69d083059e14020e626c6e70c9a34ca435da35bda27d50580b6 not found: ID does not exist" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.273345 4751 scope.go:117] "RemoveContainer" containerID="3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a" Oct 02 12:01:15 crc kubenswrapper[4751]: E1002 12:01:15.273812 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a\": container with ID starting with 3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a not found: ID does not exist" containerID="3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.273954 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a"} err="failed to get container status \"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a\": rpc error: code = NotFound desc = could not find container \"3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a\": container with ID starting with 3881ea3ed85a6fb26892fd2dc745b452c72d5ad5a7a9508b0d3883367346df4a not found: ID does not exist" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.274042 4751 scope.go:117] "RemoveContainer" containerID="24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179" Oct 02 12:01:15 crc kubenswrapper[4751]: E1002 12:01:15.274443 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179\": container with ID starting with 24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179 not found: ID does not exist" containerID="24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.275402 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179"} err="failed to get container status \"24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179\": rpc error: code = NotFound desc = could not find container \"24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179\": container with ID starting with 24aa29636c0f970cb1a023a4489fe991897be35750bbb2b626c3f914286f8179 not found: ID does not exist" Oct 02 12:01:15 crc kubenswrapper[4751]: I1002 12:01:15.561309 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" path="/var/lib/kubelet/pods/dac93481-cb79-4f37-9da8-73cc941e9a63/volumes" Oct 02 12:01:31 crc kubenswrapper[4751]: I1002 12:01:31.507054 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:01:31 crc kubenswrapper[4751]: I1002 12:01:31.507743 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.144529 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:01:47 crc kubenswrapper[4751]: E1002 12:01:47.145732 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" containerName="keystone-cron" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.145752 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" containerName="keystone-cron" Oct 02 12:01:47 crc kubenswrapper[4751]: E1002 12:01:47.145780 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="registry-server" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.145789 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="registry-server" Oct 02 12:01:47 crc kubenswrapper[4751]: E1002 12:01:47.145828 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="extract-utilities" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.145837 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="extract-utilities" Oct 02 12:01:47 crc kubenswrapper[4751]: E1002 12:01:47.145855 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="extract-content" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.145863 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="extract-content" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.146105 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96" containerName="keystone-cron" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.146133 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac93481-cb79-4f37-9da8-73cc941e9a63" containerName="registry-server" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.147954 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.177849 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lr5z\" (UniqueName: \"kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.178141 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.178260 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.178801 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.280148 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lr5z\" (UniqueName: \"kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.280419 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.280464 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.281015 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.281034 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.303391 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lr5z\" (UniqueName: \"kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z\") pod \"redhat-marketplace-2mrgg\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.481425 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:47 crc kubenswrapper[4751]: I1002 12:01:47.940730 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:01:48 crc kubenswrapper[4751]: W1002 12:01:48.363726 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5723a4bc_61b5_461d_ab24_333a0e203a75.slice/crio-e10944c1dc03478f09c8ffae3e705f849350616a67d54fb16fcceca81817faa4 WatchSource:0}: Error finding container e10944c1dc03478f09c8ffae3e705f849350616a67d54fb16fcceca81817faa4: Status 404 returned error can't find the container with id e10944c1dc03478f09c8ffae3e705f849350616a67d54fb16fcceca81817faa4 Oct 02 12:01:48 crc kubenswrapper[4751]: I1002 12:01:48.499074 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerStarted","Data":"e10944c1dc03478f09c8ffae3e705f849350616a67d54fb16fcceca81817faa4"} Oct 02 12:01:49 crc kubenswrapper[4751]: I1002 12:01:49.515763 4751 generic.go:334] "Generic (PLEG): container finished" podID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerID="a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732" exitCode=0 Oct 02 12:01:49 crc kubenswrapper[4751]: I1002 12:01:49.515875 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerDied","Data":"a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732"} Oct 02 12:01:50 crc kubenswrapper[4751]: I1002 12:01:50.529940 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerStarted","Data":"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff"} Oct 02 12:01:51 crc kubenswrapper[4751]: I1002 12:01:51.548022 4751 generic.go:334] "Generic (PLEG): container finished" podID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerID="2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff" exitCode=0 Oct 02 12:01:51 crc kubenswrapper[4751]: I1002 12:01:51.548143 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerDied","Data":"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff"} Oct 02 12:01:52 crc kubenswrapper[4751]: I1002 12:01:52.562539 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerStarted","Data":"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca"} Oct 02 12:01:52 crc kubenswrapper[4751]: I1002 12:01:52.607065 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2mrgg" podStartSLOduration=3.060952436 podStartE2EDuration="5.607032689s" podCreationTimestamp="2025-10-02 12:01:47 +0000 UTC" firstStartedPulling="2025-10-02 12:01:49.519143114 +0000 UTC m=+4191.573369604" lastFinishedPulling="2025-10-02 12:01:52.065223397 +0000 UTC m=+4194.119449857" observedRunningTime="2025-10-02 12:01:52.59262511 +0000 UTC m=+4194.646851600" watchObservedRunningTime="2025-10-02 12:01:52.607032689 +0000 UTC m=+4194.661259209" Oct 02 12:01:57 crc kubenswrapper[4751]: I1002 12:01:57.481676 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:57 crc kubenswrapper[4751]: I1002 12:01:57.482605 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:57 crc kubenswrapper[4751]: I1002 12:01:57.548502 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:57 crc kubenswrapper[4751]: I1002 12:01:57.697083 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:01:57 crc kubenswrapper[4751]: I1002 12:01:57.787706 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:01:59 crc kubenswrapper[4751]: I1002 12:01:59.634125 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2mrgg" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="registry-server" containerID="cri-o://c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca" gracePeriod=2 Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.107036 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.249120 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lr5z\" (UniqueName: \"kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z\") pod \"5723a4bc-61b5-461d-ab24-333a0e203a75\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.249310 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities\") pod \"5723a4bc-61b5-461d-ab24-333a0e203a75\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.249375 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content\") pod \"5723a4bc-61b5-461d-ab24-333a0e203a75\" (UID: \"5723a4bc-61b5-461d-ab24-333a0e203a75\") " Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.250723 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities" (OuterVolumeSpecName: "utilities") pod "5723a4bc-61b5-461d-ab24-333a0e203a75" (UID: "5723a4bc-61b5-461d-ab24-333a0e203a75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.266939 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5723a4bc-61b5-461d-ab24-333a0e203a75" (UID: "5723a4bc-61b5-461d-ab24-333a0e203a75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.352591 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.352642 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5723a4bc-61b5-461d-ab24-333a0e203a75-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.655363 4751 generic.go:334] "Generic (PLEG): container finished" podID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerID="c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca" exitCode=0 Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.655441 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerDied","Data":"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca"} Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.655489 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mrgg" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.655518 4751 scope.go:117] "RemoveContainer" containerID="c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.655501 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mrgg" event={"ID":"5723a4bc-61b5-461d-ab24-333a0e203a75","Type":"ContainerDied","Data":"e10944c1dc03478f09c8ffae3e705f849350616a67d54fb16fcceca81817faa4"} Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.684738 4751 scope.go:117] "RemoveContainer" containerID="2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.758812 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z" (OuterVolumeSpecName: "kube-api-access-2lr5z") pod "5723a4bc-61b5-461d-ab24-333a0e203a75" (UID: "5723a4bc-61b5-461d-ab24-333a0e203a75"). InnerVolumeSpecName "kube-api-access-2lr5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.762226 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lr5z\" (UniqueName: \"kubernetes.io/projected/5723a4bc-61b5-461d-ab24-333a0e203a75-kube-api-access-2lr5z\") on node \"crc\" DevicePath \"\"" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.785368 4751 scope.go:117] "RemoveContainer" containerID="a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.885888 4751 scope.go:117] "RemoveContainer" containerID="c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca" Oct 02 12:02:00 crc kubenswrapper[4751]: E1002 12:02:00.886470 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca\": container with ID starting with c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca not found: ID does not exist" containerID="c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.886566 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca"} err="failed to get container status \"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca\": rpc error: code = NotFound desc = could not find container \"c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca\": container with ID starting with c98929f7c58cd148c3a42bebc482b0d8d7c4329b9413593bd761bfa6ab960fca not found: ID does not exist" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.886626 4751 scope.go:117] "RemoveContainer" containerID="2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff" Oct 02 12:02:00 crc kubenswrapper[4751]: E1002 12:02:00.886981 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff\": container with ID starting with 2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff not found: ID does not exist" containerID="2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.887036 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff"} err="failed to get container status \"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff\": rpc error: code = NotFound desc = could not find container \"2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff\": container with ID starting with 2cccccf3950e2dbb2c45d9c6afb0bc82416de5ef2d4fc7124990d8bcd2fff2ff not found: ID does not exist" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.887077 4751 scope.go:117] "RemoveContainer" containerID="a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732" Oct 02 12:02:00 crc kubenswrapper[4751]: E1002 12:02:00.887738 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732\": container with ID starting with a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732 not found: ID does not exist" containerID="a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732" Oct 02 12:02:00 crc kubenswrapper[4751]: I1002 12:02:00.887781 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732"} err="failed to get container status \"a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732\": rpc error: code = NotFound desc = could not find container \"a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732\": container with ID starting with a995bfe556f45326ba6999c0ee0212022b6ea0956305bbce7e7b8f3780f07732 not found: ID does not exist" Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.004728 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.013921 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mrgg"] Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.506767 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.506883 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.506966 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.508379 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.508461 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e" gracePeriod=600 Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.562510 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" path="/var/lib/kubelet/pods/5723a4bc-61b5-461d-ab24-333a0e203a75/volumes" Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.671511 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e" exitCode=0 Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.671576 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e"} Oct 02 12:02:01 crc kubenswrapper[4751]: I1002 12:02:01.672004 4751 scope.go:117] "RemoveContainer" containerID="4152825dc8fff9b8a3d082ad01e4c71382dc5233ddcc5bc990d35a831156bb8f" Oct 02 12:02:02 crc kubenswrapper[4751]: I1002 12:02:02.684868 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286"} Oct 02 12:04:00 crc kubenswrapper[4751]: I1002 12:04:00.759872 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9a0ea287-4160-468d-b606-efdbc47e9c50" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 02 12:04:00 crc kubenswrapper[4751]: I1002 12:04:00.837148 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9a0ea287-4160-468d-b606-efdbc47e9c50" containerName="ceilometer-central-agent" probeResult="failure" output=< Oct 02 12:04:00 crc kubenswrapper[4751]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 02 12:04:00 crc kubenswrapper[4751]: > Oct 02 12:04:01 crc kubenswrapper[4751]: I1002 12:04:01.507750 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:04:01 crc kubenswrapper[4751]: I1002 12:04:01.508199 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.743863 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:03 crc kubenswrapper[4751]: E1002 12:04:03.744741 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="extract-utilities" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.744764 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="extract-utilities" Oct 02 12:04:03 crc kubenswrapper[4751]: E1002 12:04:03.744812 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="extract-content" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.744827 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="extract-content" Oct 02 12:04:03 crc kubenswrapper[4751]: E1002 12:04:03.744862 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="registry-server" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.744875 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="registry-server" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.745254 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5723a4bc-61b5-461d-ab24-333a0e203a75" containerName="registry-server" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.747475 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.770297 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.833549 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.833763 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fl7x\" (UniqueName: \"kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.833901 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.936017 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.936113 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fl7x\" (UniqueName: \"kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.936153 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.936716 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.936728 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:03 crc kubenswrapper[4751]: I1002 12:04:03.957258 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fl7x\" (UniqueName: \"kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x\") pod \"community-operators-d26wj\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:04 crc kubenswrapper[4751]: I1002 12:04:04.088814 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:04 crc kubenswrapper[4751]: I1002 12:04:04.923457 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:04 crc kubenswrapper[4751]: I1002 12:04:04.996227 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerStarted","Data":"fa8641097cde182491480008b14c2ab54df2efdbdb00fee5805df20720269f28"} Oct 02 12:04:05 crc kubenswrapper[4751]: I1002 12:04:05.797844 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="9a0ea287-4160-468d-b606-efdbc47e9c50" containerName="ceilometer-central-agent" probeResult="failure" output=< Oct 02 12:04:05 crc kubenswrapper[4751]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 02 12:04:05 crc kubenswrapper[4751]: > Oct 02 12:04:05 crc kubenswrapper[4751]: I1002 12:04:05.798234 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 02 12:04:05 crc kubenswrapper[4751]: I1002 12:04:05.798967 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"b921d098eaef98400b81e4dfcb4b97fac188a0f675c78762ab537dd6a3f9014a"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Oct 02 12:04:05 crc kubenswrapper[4751]: I1002 12:04:05.799086 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a0ea287-4160-468d-b606-efdbc47e9c50" containerName="ceilometer-central-agent" containerID="cri-o://b921d098eaef98400b81e4dfcb4b97fac188a0f675c78762ab537dd6a3f9014a" gracePeriod=30 Oct 02 12:04:06 crc kubenswrapper[4751]: I1002 12:04:06.009840 4751 generic.go:334] "Generic (PLEG): container finished" podID="32040baf-5e24-4844-94dc-183066941811" containerID="d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686" exitCode=0 Oct 02 12:04:06 crc kubenswrapper[4751]: I1002 12:04:06.009904 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerDied","Data":"d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686"} Oct 02 12:04:06 crc kubenswrapper[4751]: I1002 12:04:06.012435 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 12:04:07 crc kubenswrapper[4751]: I1002 12:04:07.028597 4751 generic.go:334] "Generic (PLEG): container finished" podID="9a0ea287-4160-468d-b606-efdbc47e9c50" containerID="b921d098eaef98400b81e4dfcb4b97fac188a0f675c78762ab537dd6a3f9014a" exitCode=0 Oct 02 12:04:07 crc kubenswrapper[4751]: I1002 12:04:07.028923 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerDied","Data":"b921d098eaef98400b81e4dfcb4b97fac188a0f675c78762ab537dd6a3f9014a"} Oct 02 12:04:09 crc kubenswrapper[4751]: I1002 12:04:09.054804 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a0ea287-4160-468d-b606-efdbc47e9c50","Type":"ContainerStarted","Data":"c02e12bbadba5e6a4eb51b8a376487e4d0102e2d9b73bb0ba593bdcfc0fd6259"} Oct 02 12:04:09 crc kubenswrapper[4751]: I1002 12:04:09.061526 4751 generic.go:334] "Generic (PLEG): container finished" podID="32040baf-5e24-4844-94dc-183066941811" containerID="7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548" exitCode=0 Oct 02 12:04:09 crc kubenswrapper[4751]: I1002 12:04:09.061596 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerDied","Data":"7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548"} Oct 02 12:04:12 crc kubenswrapper[4751]: I1002 12:04:12.095087 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerStarted","Data":"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988"} Oct 02 12:04:12 crc kubenswrapper[4751]: I1002 12:04:12.118609 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d26wj" podStartSLOduration=3.801063658 podStartE2EDuration="9.118583698s" podCreationTimestamp="2025-10-02 12:04:03 +0000 UTC" firstStartedPulling="2025-10-02 12:04:06.011922138 +0000 UTC m=+4328.066148628" lastFinishedPulling="2025-10-02 12:04:11.329442218 +0000 UTC m=+4333.383668668" observedRunningTime="2025-10-02 12:04:12.113751558 +0000 UTC m=+4334.167978008" watchObservedRunningTime="2025-10-02 12:04:12.118583698 +0000 UTC m=+4334.172810158" Oct 02 12:04:14 crc kubenswrapper[4751]: I1002 12:04:14.089784 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:14 crc kubenswrapper[4751]: I1002 12:04:14.090083 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:14 crc kubenswrapper[4751]: I1002 12:04:14.142583 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:16 crc kubenswrapper[4751]: I1002 12:04:16.319412 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:16 crc kubenswrapper[4751]: I1002 12:04:16.368519 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.143144 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d26wj" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="registry-server" containerID="cri-o://dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988" gracePeriod=2 Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.603919 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.614433 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities\") pod \"32040baf-5e24-4844-94dc-183066941811\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.616041 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities" (OuterVolumeSpecName: "utilities") pod "32040baf-5e24-4844-94dc-183066941811" (UID: "32040baf-5e24-4844-94dc-183066941811"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.715888 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fl7x\" (UniqueName: \"kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x\") pod \"32040baf-5e24-4844-94dc-183066941811\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.715988 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content\") pod \"32040baf-5e24-4844-94dc-183066941811\" (UID: \"32040baf-5e24-4844-94dc-183066941811\") " Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.716508 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.721377 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x" (OuterVolumeSpecName: "kube-api-access-2fl7x") pod "32040baf-5e24-4844-94dc-183066941811" (UID: "32040baf-5e24-4844-94dc-183066941811"). InnerVolumeSpecName "kube-api-access-2fl7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.771206 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32040baf-5e24-4844-94dc-183066941811" (UID: "32040baf-5e24-4844-94dc-183066941811"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.818113 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fl7x\" (UniqueName: \"kubernetes.io/projected/32040baf-5e24-4844-94dc-183066941811-kube-api-access-2fl7x\") on node \"crc\" DevicePath \"\"" Oct 02 12:04:18 crc kubenswrapper[4751]: I1002 12:04:18.818164 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32040baf-5e24-4844-94dc-183066941811-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.152839 4751 generic.go:334] "Generic (PLEG): container finished" podID="32040baf-5e24-4844-94dc-183066941811" containerID="dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988" exitCode=0 Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.153026 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerDied","Data":"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988"} Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.153643 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d26wj" event={"ID":"32040baf-5e24-4844-94dc-183066941811","Type":"ContainerDied","Data":"fa8641097cde182491480008b14c2ab54df2efdbdb00fee5805df20720269f28"} Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.153115 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d26wj" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.153726 4751 scope.go:117] "RemoveContainer" containerID="dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.181090 4751 scope.go:117] "RemoveContainer" containerID="7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.186802 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.197479 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d26wj"] Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.217055 4751 scope.go:117] "RemoveContainer" containerID="d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.242735 4751 scope.go:117] "RemoveContainer" containerID="dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988" Oct 02 12:04:19 crc kubenswrapper[4751]: E1002 12:04:19.243073 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988\": container with ID starting with dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988 not found: ID does not exist" containerID="dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.243102 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988"} err="failed to get container status \"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988\": rpc error: code = NotFound desc = could not find container \"dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988\": container with ID starting with dd2473a7a28ea99af21df7deea6469fb14edd05950f2258d44e1528a52744988 not found: ID does not exist" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.243127 4751 scope.go:117] "RemoveContainer" containerID="7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548" Oct 02 12:04:19 crc kubenswrapper[4751]: E1002 12:04:19.243635 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548\": container with ID starting with 7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548 not found: ID does not exist" containerID="7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.243658 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548"} err="failed to get container status \"7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548\": rpc error: code = NotFound desc = could not find container \"7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548\": container with ID starting with 7de8f9b8930afffe563deafb2467983505c6dd626602bb34cda39595b4a9a548 not found: ID does not exist" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.243672 4751 scope.go:117] "RemoveContainer" containerID="d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686" Oct 02 12:04:19 crc kubenswrapper[4751]: E1002 12:04:19.244164 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686\": container with ID starting with d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686 not found: ID does not exist" containerID="d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.244238 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686"} err="failed to get container status \"d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686\": rpc error: code = NotFound desc = could not find container \"d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686\": container with ID starting with d27ead109a4a778709d09e21c65a576c5102abd0c3b9f4579933d94ce02e1686 not found: ID does not exist" Oct 02 12:04:19 crc kubenswrapper[4751]: I1002 12:04:19.559738 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32040baf-5e24-4844-94dc-183066941811" path="/var/lib/kubelet/pods/32040baf-5e24-4844-94dc-183066941811/volumes" Oct 02 12:04:31 crc kubenswrapper[4751]: I1002 12:04:31.507600 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:04:31 crc kubenswrapper[4751]: I1002 12:04:31.508234 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:05:01 crc kubenswrapper[4751]: I1002 12:05:01.507639 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:05:01 crc kubenswrapper[4751]: I1002 12:05:01.508542 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:05:01 crc kubenswrapper[4751]: I1002 12:05:01.508598 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:05:01 crc kubenswrapper[4751]: I1002 12:05:01.509475 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:05:01 crc kubenswrapper[4751]: I1002 12:05:01.509549 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" gracePeriod=600 Oct 02 12:05:01 crc kubenswrapper[4751]: E1002 12:05:01.635682 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:05:02 crc kubenswrapper[4751]: I1002 12:05:02.619657 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" exitCode=0 Oct 02 12:05:02 crc kubenswrapper[4751]: I1002 12:05:02.619747 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286"} Oct 02 12:05:02 crc kubenswrapper[4751]: I1002 12:05:02.619827 4751 scope.go:117] "RemoveContainer" containerID="d9d9ca816c670c78cf0bbe8df501e5bb6adfeab3b4f899c2782eddf27e2ced1e" Oct 02 12:05:02 crc kubenswrapper[4751]: I1002 12:05:02.620686 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:05:02 crc kubenswrapper[4751]: E1002 12:05:02.621155 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:05:17 crc kubenswrapper[4751]: I1002 12:05:17.551117 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:05:17 crc kubenswrapper[4751]: E1002 12:05:17.551991 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:05:30 crc kubenswrapper[4751]: I1002 12:05:30.550662 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:05:30 crc kubenswrapper[4751]: E1002 12:05:30.551744 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:05:41 crc kubenswrapper[4751]: I1002 12:05:41.550435 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:05:41 crc kubenswrapper[4751]: E1002 12:05:41.551954 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:05:53 crc kubenswrapper[4751]: I1002 12:05:53.550509 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:05:53 crc kubenswrapper[4751]: E1002 12:05:53.551443 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:06:08 crc kubenswrapper[4751]: I1002 12:06:08.550892 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:06:08 crc kubenswrapper[4751]: E1002 12:06:08.551710 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:06:20 crc kubenswrapper[4751]: I1002 12:06:20.550948 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:06:20 crc kubenswrapper[4751]: E1002 12:06:20.552464 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:06:34 crc kubenswrapper[4751]: I1002 12:06:34.549938 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:06:34 crc kubenswrapper[4751]: E1002 12:06:34.550617 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:06:48 crc kubenswrapper[4751]: I1002 12:06:48.557480 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:06:48 crc kubenswrapper[4751]: E1002 12:06:48.558563 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:07:02 crc kubenswrapper[4751]: I1002 12:07:02.549868 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:07:02 crc kubenswrapper[4751]: E1002 12:07:02.550802 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:07:16 crc kubenswrapper[4751]: I1002 12:07:16.549717 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:07:16 crc kubenswrapper[4751]: E1002 12:07:16.550384 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:07:30 crc kubenswrapper[4751]: I1002 12:07:30.551324 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:07:30 crc kubenswrapper[4751]: E1002 12:07:30.552227 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:07:41 crc kubenswrapper[4751]: I1002 12:07:41.550781 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:07:41 crc kubenswrapper[4751]: E1002 12:07:41.551760 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:07:55 crc kubenswrapper[4751]: I1002 12:07:55.552025 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:07:55 crc kubenswrapper[4751]: E1002 12:07:55.553230 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:08:10 crc kubenswrapper[4751]: I1002 12:08:10.550999 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:08:10 crc kubenswrapper[4751]: E1002 12:08:10.552227 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:08:25 crc kubenswrapper[4751]: I1002 12:08:25.550874 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:08:25 crc kubenswrapper[4751]: E1002 12:08:25.551647 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:08:37 crc kubenswrapper[4751]: I1002 12:08:37.550555 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:08:37 crc kubenswrapper[4751]: E1002 12:08:37.551448 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:08:51 crc kubenswrapper[4751]: I1002 12:08:51.550936 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:08:51 crc kubenswrapper[4751]: E1002 12:08:51.552310 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:04 crc kubenswrapper[4751]: I1002 12:09:04.551098 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:09:04 crc kubenswrapper[4751]: E1002 12:09:04.551919 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:18 crc kubenswrapper[4751]: I1002 12:09:18.550471 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:09:18 crc kubenswrapper[4751]: E1002 12:09:18.551282 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:29 crc kubenswrapper[4751]: I1002 12:09:29.556139 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:09:29 crc kubenswrapper[4751]: E1002 12:09:29.557548 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.702332 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:40 crc kubenswrapper[4751]: E1002 12:09:40.703330 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="extract-utilities" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.703346 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="extract-utilities" Oct 02 12:09:40 crc kubenswrapper[4751]: E1002 12:09:40.703373 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="registry-server" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.703381 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="registry-server" Oct 02 12:09:40 crc kubenswrapper[4751]: E1002 12:09:40.703405 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="extract-content" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.703413 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="extract-content" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.703655 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="32040baf-5e24-4844-94dc-183066941811" containerName="registry-server" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.705303 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.716737 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.822225 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.822274 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxjn\" (UniqueName: \"kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.822309 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.924327 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.924545 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.924578 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxjn\" (UniqueName: \"kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.924749 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.924952 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:40 crc kubenswrapper[4751]: I1002 12:09:40.972153 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxjn\" (UniqueName: \"kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn\") pod \"certified-operators-j6g9t\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:41 crc kubenswrapper[4751]: I1002 12:09:41.037793 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:41 crc kubenswrapper[4751]: I1002 12:09:41.518617 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:41 crc kubenswrapper[4751]: I1002 12:09:41.552706 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:09:41 crc kubenswrapper[4751]: E1002 12:09:41.553372 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:41 crc kubenswrapper[4751]: I1002 12:09:41.586364 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerStarted","Data":"b682eee83a050a61972fbc19500f1f434741bb1af6a5d89314419f9671f124ea"} Oct 02 12:09:42 crc kubenswrapper[4751]: I1002 12:09:42.599490 4751 generic.go:334] "Generic (PLEG): container finished" podID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerID="2713c6e79b16a5f7844199317264034267964f5dfb45e1a0de949ed30d29e753" exitCode=0 Oct 02 12:09:42 crc kubenswrapper[4751]: I1002 12:09:42.599547 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerDied","Data":"2713c6e79b16a5f7844199317264034267964f5dfb45e1a0de949ed30d29e753"} Oct 02 12:09:42 crc kubenswrapper[4751]: I1002 12:09:42.601818 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 12:09:44 crc kubenswrapper[4751]: I1002 12:09:44.624302 4751 generic.go:334] "Generic (PLEG): container finished" podID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerID="84c296c1d7550dd4dad71c517032a72189a40124a9d29b24d0d2fb1f1b83accb" exitCode=0 Oct 02 12:09:44 crc kubenswrapper[4751]: I1002 12:09:44.624422 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerDied","Data":"84c296c1d7550dd4dad71c517032a72189a40124a9d29b24d0d2fb1f1b83accb"} Oct 02 12:09:46 crc kubenswrapper[4751]: I1002 12:09:46.653808 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerStarted","Data":"f1636bce2edc0bb10f5e9bb4894091ec4ba1658dd730d2aa6251464ddd6f736c"} Oct 02 12:09:46 crc kubenswrapper[4751]: I1002 12:09:46.682888 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j6g9t" podStartSLOduration=3.890468858 podStartE2EDuration="6.682871063s" podCreationTimestamp="2025-10-02 12:09:40 +0000 UTC" firstStartedPulling="2025-10-02 12:09:42.601458915 +0000 UTC m=+4664.655685375" lastFinishedPulling="2025-10-02 12:09:45.39386113 +0000 UTC m=+4667.448087580" observedRunningTime="2025-10-02 12:09:46.68278517 +0000 UTC m=+4668.737011690" watchObservedRunningTime="2025-10-02 12:09:46.682871063 +0000 UTC m=+4668.737097523" Oct 02 12:09:51 crc kubenswrapper[4751]: I1002 12:09:51.037969 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:51 crc kubenswrapper[4751]: I1002 12:09:51.039341 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:51 crc kubenswrapper[4751]: I1002 12:09:51.128020 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:51 crc kubenswrapper[4751]: I1002 12:09:51.782151 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:51 crc kubenswrapper[4751]: I1002 12:09:51.831384 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:53 crc kubenswrapper[4751]: I1002 12:09:53.729297 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j6g9t" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="registry-server" containerID="cri-o://f1636bce2edc0bb10f5e9bb4894091ec4ba1658dd730d2aa6251464ddd6f736c" gracePeriod=2 Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.551082 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:09:54 crc kubenswrapper[4751]: E1002 12:09:54.552018 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.741930 4751 generic.go:334] "Generic (PLEG): container finished" podID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerID="f1636bce2edc0bb10f5e9bb4894091ec4ba1658dd730d2aa6251464ddd6f736c" exitCode=0 Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.742005 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerDied","Data":"f1636bce2edc0bb10f5e9bb4894091ec4ba1658dd730d2aa6251464ddd6f736c"} Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.867890 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.918103 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities\") pod \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.918518 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content\") pod \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.918655 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hxjn\" (UniqueName: \"kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn\") pod \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\" (UID: \"0b328844-f8bf-4c6d-98c8-2e81d04e856f\") " Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.919279 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities" (OuterVolumeSpecName: "utilities") pod "0b328844-f8bf-4c6d-98c8-2e81d04e856f" (UID: "0b328844-f8bf-4c6d-98c8-2e81d04e856f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:09:54 crc kubenswrapper[4751]: I1002 12:09:54.933502 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn" (OuterVolumeSpecName: "kube-api-access-4hxjn") pod "0b328844-f8bf-4c6d-98c8-2e81d04e856f" (UID: "0b328844-f8bf-4c6d-98c8-2e81d04e856f"). InnerVolumeSpecName "kube-api-access-4hxjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.022809 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.022845 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hxjn\" (UniqueName: \"kubernetes.io/projected/0b328844-f8bf-4c6d-98c8-2e81d04e856f-kube-api-access-4hxjn\") on node \"crc\" DevicePath \"\"" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.048854 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b328844-f8bf-4c6d-98c8-2e81d04e856f" (UID: "0b328844-f8bf-4c6d-98c8-2e81d04e856f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.124699 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b328844-f8bf-4c6d-98c8-2e81d04e856f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.754755 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6g9t" event={"ID":"0b328844-f8bf-4c6d-98c8-2e81d04e856f","Type":"ContainerDied","Data":"b682eee83a050a61972fbc19500f1f434741bb1af6a5d89314419f9671f124ea"} Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.755091 4751 scope.go:117] "RemoveContainer" containerID="f1636bce2edc0bb10f5e9bb4894091ec4ba1658dd730d2aa6251464ddd6f736c" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.754863 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6g9t" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.780552 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.791478 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j6g9t"] Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.795390 4751 scope.go:117] "RemoveContainer" containerID="84c296c1d7550dd4dad71c517032a72189a40124a9d29b24d0d2fb1f1b83accb" Oct 02 12:09:55 crc kubenswrapper[4751]: I1002 12:09:55.836367 4751 scope.go:117] "RemoveContainer" containerID="2713c6e79b16a5f7844199317264034267964f5dfb45e1a0de949ed30d29e753" Oct 02 12:09:57 crc kubenswrapper[4751]: I1002 12:09:57.563115 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" path="/var/lib/kubelet/pods/0b328844-f8bf-4c6d-98c8-2e81d04e856f/volumes" Oct 02 12:10:08 crc kubenswrapper[4751]: I1002 12:10:08.551301 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:10:08 crc kubenswrapper[4751]: I1002 12:10:08.884114 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d"} Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.990297 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:11:56 crc kubenswrapper[4751]: E1002 12:11:56.991792 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="extract-content" Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.991809 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="extract-content" Oct 02 12:11:56 crc kubenswrapper[4751]: E1002 12:11:56.991831 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="extract-utilities" Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.991840 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="extract-utilities" Oct 02 12:11:56 crc kubenswrapper[4751]: E1002 12:11:56.991874 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="registry-server" Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.991881 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="registry-server" Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.992112 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b328844-f8bf-4c6d-98c8-2e81d04e856f" containerName="registry-server" Oct 02 12:11:56 crc kubenswrapper[4751]: I1002 12:11:56.993861 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.012645 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.116750 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.116921 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v22bm\" (UniqueName: \"kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.117083 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.218627 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.218749 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.218800 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v22bm\" (UniqueName: \"kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.219274 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.219507 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.252293 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v22bm\" (UniqueName: \"kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm\") pod \"redhat-marketplace-g9j9n\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.330638 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:11:57 crc kubenswrapper[4751]: I1002 12:11:57.801108 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:11:58 crc kubenswrapper[4751]: I1002 12:11:58.073921 4751 generic.go:334] "Generic (PLEG): container finished" podID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerID="d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0" exitCode=0 Oct 02 12:11:58 crc kubenswrapper[4751]: I1002 12:11:58.073995 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerDied","Data":"d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0"} Oct 02 12:11:58 crc kubenswrapper[4751]: I1002 12:11:58.075009 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerStarted","Data":"6177207cb8a478dd4cac9395400159e8f38a2f29c30202249809d98d296e4358"} Oct 02 12:11:59 crc kubenswrapper[4751]: I1002 12:11:59.086909 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerStarted","Data":"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60"} Oct 02 12:12:00 crc kubenswrapper[4751]: I1002 12:12:00.099479 4751 generic.go:334] "Generic (PLEG): container finished" podID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerID="5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60" exitCode=0 Oct 02 12:12:00 crc kubenswrapper[4751]: I1002 12:12:00.099515 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerDied","Data":"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60"} Oct 02 12:12:01 crc kubenswrapper[4751]: I1002 12:12:01.109388 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerStarted","Data":"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda"} Oct 02 12:12:01 crc kubenswrapper[4751]: I1002 12:12:01.128995 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g9j9n" podStartSLOduration=2.722365339 podStartE2EDuration="5.128972591s" podCreationTimestamp="2025-10-02 12:11:56 +0000 UTC" firstStartedPulling="2025-10-02 12:11:58.075875419 +0000 UTC m=+4800.130101879" lastFinishedPulling="2025-10-02 12:12:00.482482681 +0000 UTC m=+4802.536709131" observedRunningTime="2025-10-02 12:12:01.125892378 +0000 UTC m=+4803.180118908" watchObservedRunningTime="2025-10-02 12:12:01.128972591 +0000 UTC m=+4803.183199041" Oct 02 12:12:07 crc kubenswrapper[4751]: I1002 12:12:07.330924 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:07 crc kubenswrapper[4751]: I1002 12:12:07.332208 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:07 crc kubenswrapper[4751]: I1002 12:12:07.401517 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:08 crc kubenswrapper[4751]: I1002 12:12:08.251060 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:08 crc kubenswrapper[4751]: I1002 12:12:08.323578 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.207997 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g9j9n" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="registry-server" containerID="cri-o://4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda" gracePeriod=2 Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.694126 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.801126 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities\") pod \"9103b791-59ed-4f35-8721-c30f64e0c9e9\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.801424 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content\") pod \"9103b791-59ed-4f35-8721-c30f64e0c9e9\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.801553 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v22bm\" (UniqueName: \"kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm\") pod \"9103b791-59ed-4f35-8721-c30f64e0c9e9\" (UID: \"9103b791-59ed-4f35-8721-c30f64e0c9e9\") " Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.801953 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities" (OuterVolumeSpecName: "utilities") pod "9103b791-59ed-4f35-8721-c30f64e0c9e9" (UID: "9103b791-59ed-4f35-8721-c30f64e0c9e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.802077 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.807695 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm" (OuterVolumeSpecName: "kube-api-access-v22bm") pod "9103b791-59ed-4f35-8721-c30f64e0c9e9" (UID: "9103b791-59ed-4f35-8721-c30f64e0c9e9"). InnerVolumeSpecName "kube-api-access-v22bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.814784 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9103b791-59ed-4f35-8721-c30f64e0c9e9" (UID: "9103b791-59ed-4f35-8721-c30f64e0c9e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.903911 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103b791-59ed-4f35-8721-c30f64e0c9e9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:10 crc kubenswrapper[4751]: I1002 12:12:10.903944 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v22bm\" (UniqueName: \"kubernetes.io/projected/9103b791-59ed-4f35-8721-c30f64e0c9e9-kube-api-access-v22bm\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.219921 4751 generic.go:334] "Generic (PLEG): container finished" podID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerID="4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda" exitCode=0 Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.219966 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerDied","Data":"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda"} Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.220228 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g9j9n" event={"ID":"9103b791-59ed-4f35-8721-c30f64e0c9e9","Type":"ContainerDied","Data":"6177207cb8a478dd4cac9395400159e8f38a2f29c30202249809d98d296e4358"} Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.220253 4751 scope.go:117] "RemoveContainer" containerID="4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.220002 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g9j9n" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.268120 4751 scope.go:117] "RemoveContainer" containerID="5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.270164 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.281439 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g9j9n"] Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.294113 4751 scope.go:117] "RemoveContainer" containerID="d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.351978 4751 scope.go:117] "RemoveContainer" containerID="4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda" Oct 02 12:12:11 crc kubenswrapper[4751]: E1002 12:12:11.352452 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda\": container with ID starting with 4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda not found: ID does not exist" containerID="4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.352485 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda"} err="failed to get container status \"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda\": rpc error: code = NotFound desc = could not find container \"4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda\": container with ID starting with 4efda74344ab09a9de1e797ab1684ea2bd3536f1945d9de37f84410adb66adda not found: ID does not exist" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.352505 4751 scope.go:117] "RemoveContainer" containerID="5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60" Oct 02 12:12:11 crc kubenswrapper[4751]: E1002 12:12:11.352804 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60\": container with ID starting with 5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60 not found: ID does not exist" containerID="5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.352826 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60"} err="failed to get container status \"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60\": rpc error: code = NotFound desc = could not find container \"5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60\": container with ID starting with 5f672719694a3a253d450b56076c969e8e7a893921d74fd91db047fb58913c60 not found: ID does not exist" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.352838 4751 scope.go:117] "RemoveContainer" containerID="d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0" Oct 02 12:12:11 crc kubenswrapper[4751]: E1002 12:12:11.353054 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0\": container with ID starting with d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0 not found: ID does not exist" containerID="d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.353077 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0"} err="failed to get container status \"d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0\": rpc error: code = NotFound desc = could not find container \"d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0\": container with ID starting with d23029301aea22c90dd29e1b91521ccfaf0afbc51440a926f7f7fafcbad3fcb0 not found: ID does not exist" Oct 02 12:12:11 crc kubenswrapper[4751]: I1002 12:12:11.566035 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" path="/var/lib/kubelet/pods/9103b791-59ed-4f35-8721-c30f64e0c9e9/volumes" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.786845 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:25 crc kubenswrapper[4751]: E1002 12:12:25.788059 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="extract-utilities" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.788082 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="extract-utilities" Oct 02 12:12:25 crc kubenswrapper[4751]: E1002 12:12:25.788128 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="extract-content" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.788141 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="extract-content" Oct 02 12:12:25 crc kubenswrapper[4751]: E1002 12:12:25.788207 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="registry-server" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.788220 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="registry-server" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.788550 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="9103b791-59ed-4f35-8721-c30f64e0c9e9" containerName="registry-server" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.790992 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.796903 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.912452 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.912498 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:25 crc kubenswrapper[4751]: I1002 12:12:25.912951 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mfmq\" (UniqueName: \"kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.014496 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mfmq\" (UniqueName: \"kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.014576 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.014592 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.015138 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.015339 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.258995 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mfmq\" (UniqueName: \"kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq\") pod \"redhat-operators-wvsx5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.428344 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:26 crc kubenswrapper[4751]: I1002 12:12:26.924532 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:27 crc kubenswrapper[4751]: I1002 12:12:27.417657 4751 generic.go:334] "Generic (PLEG): container finished" podID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerID="9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f" exitCode=0 Oct 02 12:12:27 crc kubenswrapper[4751]: I1002 12:12:27.417736 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerDied","Data":"9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f"} Oct 02 12:12:27 crc kubenswrapper[4751]: I1002 12:12:27.417895 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerStarted","Data":"c690c8965ee48f4d0e4f5c5ef125f53f8bf1724af112b368b9a5848bbd58c18c"} Oct 02 12:12:29 crc kubenswrapper[4751]: I1002 12:12:29.441489 4751 generic.go:334] "Generic (PLEG): container finished" podID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerID="36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3" exitCode=0 Oct 02 12:12:29 crc kubenswrapper[4751]: I1002 12:12:29.441609 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerDied","Data":"36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3"} Oct 02 12:12:30 crc kubenswrapper[4751]: I1002 12:12:30.453127 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerStarted","Data":"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c"} Oct 02 12:12:30 crc kubenswrapper[4751]: I1002 12:12:30.481417 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wvsx5" podStartSLOduration=2.90755922 podStartE2EDuration="5.481396983s" podCreationTimestamp="2025-10-02 12:12:25 +0000 UTC" firstStartedPulling="2025-10-02 12:12:27.419934386 +0000 UTC m=+4829.474160876" lastFinishedPulling="2025-10-02 12:12:29.993772189 +0000 UTC m=+4832.047998639" observedRunningTime="2025-10-02 12:12:30.473079729 +0000 UTC m=+4832.527306179" watchObservedRunningTime="2025-10-02 12:12:30.481396983 +0000 UTC m=+4832.535623443" Oct 02 12:12:31 crc kubenswrapper[4751]: I1002 12:12:31.507094 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:12:31 crc kubenswrapper[4751]: I1002 12:12:31.507519 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:12:36 crc kubenswrapper[4751]: I1002 12:12:36.430282 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:36 crc kubenswrapper[4751]: I1002 12:12:36.431222 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:36 crc kubenswrapper[4751]: I1002 12:12:36.494888 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:36 crc kubenswrapper[4751]: I1002 12:12:36.577105 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:36 crc kubenswrapper[4751]: I1002 12:12:36.738599 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:38 crc kubenswrapper[4751]: I1002 12:12:38.544683 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wvsx5" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="registry-server" containerID="cri-o://9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c" gracePeriod=2 Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.000267 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.104489 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities\") pod \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.104571 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mfmq\" (UniqueName: \"kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq\") pod \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.104710 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content\") pod \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\" (UID: \"136b63b1-c47a-4d61-82ff-b5c37f674bc5\") " Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.105216 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities" (OuterVolumeSpecName: "utilities") pod "136b63b1-c47a-4d61-82ff-b5c37f674bc5" (UID: "136b63b1-c47a-4d61-82ff-b5c37f674bc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.206732 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.565978 4751 generic.go:334] "Generic (PLEG): container finished" podID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerID="9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c" exitCode=0 Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.566117 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wvsx5" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.568953 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerDied","Data":"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c"} Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.568993 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wvsx5" event={"ID":"136b63b1-c47a-4d61-82ff-b5c37f674bc5","Type":"ContainerDied","Data":"c690c8965ee48f4d0e4f5c5ef125f53f8bf1724af112b368b9a5848bbd58c18c"} Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.569018 4751 scope.go:117] "RemoveContainer" containerID="9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.602312 4751 scope.go:117] "RemoveContainer" containerID="36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.657578 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq" (OuterVolumeSpecName: "kube-api-access-7mfmq") pod "136b63b1-c47a-4d61-82ff-b5c37f674bc5" (UID: "136b63b1-c47a-4d61-82ff-b5c37f674bc5"). InnerVolumeSpecName "kube-api-access-7mfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.681556 4751 scope.go:117] "RemoveContainer" containerID="9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.717200 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mfmq\" (UniqueName: \"kubernetes.io/projected/136b63b1-c47a-4d61-82ff-b5c37f674bc5-kube-api-access-7mfmq\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.871997 4751 scope.go:117] "RemoveContainer" containerID="9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c" Oct 02 12:12:39 crc kubenswrapper[4751]: E1002 12:12:39.872411 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c\": container with ID starting with 9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c not found: ID does not exist" containerID="9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.872442 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c"} err="failed to get container status \"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c\": rpc error: code = NotFound desc = could not find container \"9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c\": container with ID starting with 9018d32afeef3fa9a52232bdc49e4b7094c07a1111410f04b4a11a9b9fe7f48c not found: ID does not exist" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.872464 4751 scope.go:117] "RemoveContainer" containerID="36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3" Oct 02 12:12:39 crc kubenswrapper[4751]: E1002 12:12:39.872732 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3\": container with ID starting with 36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3 not found: ID does not exist" containerID="36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.872754 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3"} err="failed to get container status \"36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3\": rpc error: code = NotFound desc = could not find container \"36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3\": container with ID starting with 36c7715070bdadec3e9ec3bee9329c72145f48bb5ce0681b6f23b2bc5bbca4f3 not found: ID does not exist" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.872767 4751 scope.go:117] "RemoveContainer" containerID="9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f" Oct 02 12:12:39 crc kubenswrapper[4751]: E1002 12:12:39.873047 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f\": container with ID starting with 9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f not found: ID does not exist" containerID="9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f" Oct 02 12:12:39 crc kubenswrapper[4751]: I1002 12:12:39.873150 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f"} err="failed to get container status \"9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f\": rpc error: code = NotFound desc = could not find container \"9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f\": container with ID starting with 9b9d0bad689b6557dac3c586ecb16f3c182674572606da68a6a2411dd26cf80f not found: ID does not exist" Oct 02 12:12:40 crc kubenswrapper[4751]: I1002 12:12:40.624405 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "136b63b1-c47a-4d61-82ff-b5c37f674bc5" (UID: "136b63b1-c47a-4d61-82ff-b5c37f674bc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:12:40 crc kubenswrapper[4751]: I1002 12:12:40.635586 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136b63b1-c47a-4d61-82ff-b5c37f674bc5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:12:40 crc kubenswrapper[4751]: I1002 12:12:40.817827 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:40 crc kubenswrapper[4751]: I1002 12:12:40.826687 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wvsx5"] Oct 02 12:12:41 crc kubenswrapper[4751]: I1002 12:12:41.563581 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" path="/var/lib/kubelet/pods/136b63b1-c47a-4d61-82ff-b5c37f674bc5/volumes" Oct 02 12:13:01 crc kubenswrapper[4751]: I1002 12:13:01.507236 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:13:01 crc kubenswrapper[4751]: I1002 12:13:01.507934 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:13:31 crc kubenswrapper[4751]: I1002 12:13:31.507420 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:13:31 crc kubenswrapper[4751]: I1002 12:13:31.507995 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:13:31 crc kubenswrapper[4751]: I1002 12:13:31.508049 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:13:31 crc kubenswrapper[4751]: I1002 12:13:31.508848 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:13:31 crc kubenswrapper[4751]: I1002 12:13:31.508922 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d" gracePeriod=600 Oct 02 12:13:32 crc kubenswrapper[4751]: I1002 12:13:32.118785 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d" exitCode=0 Oct 02 12:13:32 crc kubenswrapper[4751]: I1002 12:13:32.118831 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d"} Oct 02 12:13:32 crc kubenswrapper[4751]: I1002 12:13:32.119158 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4"} Oct 02 12:13:32 crc kubenswrapper[4751]: I1002 12:13:32.119201 4751 scope.go:117] "RemoveContainer" containerID="d374931498ac6f81947b59a77edd9dc2318628f58d43790b207fda957bf23286" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.302097 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:09 crc kubenswrapper[4751]: E1002 12:14:09.303101 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="extract-utilities" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.303115 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="extract-utilities" Oct 02 12:14:09 crc kubenswrapper[4751]: E1002 12:14:09.303129 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="registry-server" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.303137 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="registry-server" Oct 02 12:14:09 crc kubenswrapper[4751]: E1002 12:14:09.303154 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="extract-content" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.303160 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="extract-content" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.303356 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="136b63b1-c47a-4d61-82ff-b5c37f674bc5" containerName="registry-server" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.304922 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.310997 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.434243 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2wx8\" (UniqueName: \"kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.434529 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.434785 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.536693 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.536775 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2wx8\" (UniqueName: \"kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.536847 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.537455 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.537475 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.564675 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2wx8\" (UniqueName: \"kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8\") pod \"community-operators-v25vh\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:09 crc kubenswrapper[4751]: I1002 12:14:09.633842 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:10 crc kubenswrapper[4751]: I1002 12:14:10.150407 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:10 crc kubenswrapper[4751]: W1002 12:14:10.159153 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5584ab5a_311b_46fe_aeed_9b7e5a548463.slice/crio-403e59d6f2c29b845f23f16992247effd1a28d39bff9b46778547521b47a4e51 WatchSource:0}: Error finding container 403e59d6f2c29b845f23f16992247effd1a28d39bff9b46778547521b47a4e51: Status 404 returned error can't find the container with id 403e59d6f2c29b845f23f16992247effd1a28d39bff9b46778547521b47a4e51 Oct 02 12:14:10 crc kubenswrapper[4751]: I1002 12:14:10.540579 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerStarted","Data":"403e59d6f2c29b845f23f16992247effd1a28d39bff9b46778547521b47a4e51"} Oct 02 12:14:10 crc kubenswrapper[4751]: E1002 12:14:10.690756 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5584ab5a_311b_46fe_aeed_9b7e5a548463.slice/crio-conmon-d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5584ab5a_311b_46fe_aeed_9b7e5a548463.slice/crio-d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077.scope\": RecentStats: unable to find data in memory cache]" Oct 02 12:14:11 crc kubenswrapper[4751]: I1002 12:14:11.551094 4751 generic.go:334] "Generic (PLEG): container finished" podID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerID="d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077" exitCode=0 Oct 02 12:14:11 crc kubenswrapper[4751]: I1002 12:14:11.563189 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerDied","Data":"d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077"} Oct 02 12:14:13 crc kubenswrapper[4751]: I1002 12:14:13.571086 4751 generic.go:334] "Generic (PLEG): container finished" podID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerID="b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2" exitCode=0 Oct 02 12:14:13 crc kubenswrapper[4751]: I1002 12:14:13.571345 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerDied","Data":"b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2"} Oct 02 12:14:15 crc kubenswrapper[4751]: I1002 12:14:15.594649 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerStarted","Data":"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710"} Oct 02 12:14:15 crc kubenswrapper[4751]: I1002 12:14:15.618064 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v25vh" podStartSLOduration=4.130843395 podStartE2EDuration="6.618043407s" podCreationTimestamp="2025-10-02 12:14:09 +0000 UTC" firstStartedPulling="2025-10-02 12:14:11.553691998 +0000 UTC m=+4933.607918448" lastFinishedPulling="2025-10-02 12:14:14.04089201 +0000 UTC m=+4936.095118460" observedRunningTime="2025-10-02 12:14:15.611716216 +0000 UTC m=+4937.665942666" watchObservedRunningTime="2025-10-02 12:14:15.618043407 +0000 UTC m=+4937.672269857" Oct 02 12:14:19 crc kubenswrapper[4751]: I1002 12:14:19.634438 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:19 crc kubenswrapper[4751]: I1002 12:14:19.635048 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:19 crc kubenswrapper[4751]: I1002 12:14:19.704115 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:19 crc kubenswrapper[4751]: I1002 12:14:19.759888 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:19 crc kubenswrapper[4751]: I1002 12:14:19.942725 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:21 crc kubenswrapper[4751]: I1002 12:14:21.687466 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v25vh" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="registry-server" containerID="cri-o://d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710" gracePeriod=2 Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.204477 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.325085 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content\") pod \"5584ab5a-311b-46fe-aeed-9b7e5a548463\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.325578 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities\") pod \"5584ab5a-311b-46fe-aeed-9b7e5a548463\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.325752 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2wx8\" (UniqueName: \"kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8\") pod \"5584ab5a-311b-46fe-aeed-9b7e5a548463\" (UID: \"5584ab5a-311b-46fe-aeed-9b7e5a548463\") " Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.326620 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities" (OuterVolumeSpecName: "utilities") pod "5584ab5a-311b-46fe-aeed-9b7e5a548463" (UID: "5584ab5a-311b-46fe-aeed-9b7e5a548463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.330907 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8" (OuterVolumeSpecName: "kube-api-access-f2wx8") pod "5584ab5a-311b-46fe-aeed-9b7e5a548463" (UID: "5584ab5a-311b-46fe-aeed-9b7e5a548463"). InnerVolumeSpecName "kube-api-access-f2wx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.387540 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5584ab5a-311b-46fe-aeed-9b7e5a548463" (UID: "5584ab5a-311b-46fe-aeed-9b7e5a548463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.429024 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.429073 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5584ab5a-311b-46fe-aeed-9b7e5a548463-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.429085 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2wx8\" (UniqueName: \"kubernetes.io/projected/5584ab5a-311b-46fe-aeed-9b7e5a548463-kube-api-access-f2wx8\") on node \"crc\" DevicePath \"\"" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.701614 4751 generic.go:334] "Generic (PLEG): container finished" podID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerID="d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710" exitCode=0 Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.701664 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerDied","Data":"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710"} Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.701692 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v25vh" event={"ID":"5584ab5a-311b-46fe-aeed-9b7e5a548463","Type":"ContainerDied","Data":"403e59d6f2c29b845f23f16992247effd1a28d39bff9b46778547521b47a4e51"} Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.701716 4751 scope.go:117] "RemoveContainer" containerID="d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.701721 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v25vh" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.740972 4751 scope.go:117] "RemoveContainer" containerID="b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.745624 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.755468 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v25vh"] Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.775422 4751 scope.go:117] "RemoveContainer" containerID="d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.806368 4751 scope.go:117] "RemoveContainer" containerID="d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710" Oct 02 12:14:22 crc kubenswrapper[4751]: E1002 12:14:22.806898 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710\": container with ID starting with d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710 not found: ID does not exist" containerID="d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.806951 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710"} err="failed to get container status \"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710\": rpc error: code = NotFound desc = could not find container \"d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710\": container with ID starting with d9f91530c6701eae92466542bbca07207597220cf4b8b5646f0a44e431d0a710 not found: ID does not exist" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.806989 4751 scope.go:117] "RemoveContainer" containerID="b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2" Oct 02 12:14:22 crc kubenswrapper[4751]: E1002 12:14:22.807438 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2\": container with ID starting with b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2 not found: ID does not exist" containerID="b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.807469 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2"} err="failed to get container status \"b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2\": rpc error: code = NotFound desc = could not find container \"b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2\": container with ID starting with b2452269f1280cc151f80acdbc6a352ae9774d7655c3bb7ad3ea3554299bf0a2 not found: ID does not exist" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.807489 4751 scope.go:117] "RemoveContainer" containerID="d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077" Oct 02 12:14:22 crc kubenswrapper[4751]: E1002 12:14:22.807755 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077\": container with ID starting with d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077 not found: ID does not exist" containerID="d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077" Oct 02 12:14:22 crc kubenswrapper[4751]: I1002 12:14:22.807775 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077"} err="failed to get container status \"d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077\": rpc error: code = NotFound desc = could not find container \"d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077\": container with ID starting with d20dfc36472c6ebb76095f643560c3cee757850ed8b6d32d2da3205eec5c0077 not found: ID does not exist" Oct 02 12:14:23 crc kubenswrapper[4751]: I1002 12:14:23.566251 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" path="/var/lib/kubelet/pods/5584ab5a-311b-46fe-aeed-9b7e5a548463/volumes" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.142793 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t"] Oct 02 12:15:00 crc kubenswrapper[4751]: E1002 12:15:00.143747 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="registry-server" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.143764 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="registry-server" Oct 02 12:15:00 crc kubenswrapper[4751]: E1002 12:15:00.143781 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="extract-content" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.143788 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="extract-content" Oct 02 12:15:00 crc kubenswrapper[4751]: E1002 12:15:00.143804 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="extract-utilities" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.143810 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="extract-utilities" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.143991 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="5584ab5a-311b-46fe-aeed-9b7e5a548463" containerName="registry-server" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.144607 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.146357 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.146632 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.150849 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t"] Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.181554 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.181717 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2jzb\" (UniqueName: \"kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.181902 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.284195 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2jzb\" (UniqueName: \"kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.284300 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.284418 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.285300 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.290308 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.301068 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2jzb\" (UniqueName: \"kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb\") pod \"collect-profiles-29323455-dkd7t\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.471042 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:00 crc kubenswrapper[4751]: W1002 12:15:00.896541 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd68a88b7_2660_43b8_8a32_512eeefb5e66.slice/crio-bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73 WatchSource:0}: Error finding container bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73: Status 404 returned error can't find the container with id bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73 Oct 02 12:15:00 crc kubenswrapper[4751]: I1002 12:15:00.898377 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t"] Oct 02 12:15:01 crc kubenswrapper[4751]: I1002 12:15:01.099334 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" event={"ID":"d68a88b7-2660-43b8-8a32-512eeefb5e66","Type":"ContainerStarted","Data":"8638a7e82c89802b5cbedd24be5e686a5cd0a38f27f1ab2844ab8b304c8f5602"} Oct 02 12:15:01 crc kubenswrapper[4751]: I1002 12:15:01.099386 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" event={"ID":"d68a88b7-2660-43b8-8a32-512eeefb5e66","Type":"ContainerStarted","Data":"bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73"} Oct 02 12:15:01 crc kubenswrapper[4751]: I1002 12:15:01.123601 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" podStartSLOduration=1.12358458 podStartE2EDuration="1.12358458s" podCreationTimestamp="2025-10-02 12:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 12:15:01.115841661 +0000 UTC m=+4983.170068111" watchObservedRunningTime="2025-10-02 12:15:01.12358458 +0000 UTC m=+4983.177811030" Oct 02 12:15:02 crc kubenswrapper[4751]: E1002 12:15:02.032754 4751 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd68a88b7_2660_43b8_8a32_512eeefb5e66.slice/crio-8638a7e82c89802b5cbedd24be5e686a5cd0a38f27f1ab2844ab8b304c8f5602.scope\": RecentStats: unable to find data in memory cache]" Oct 02 12:15:02 crc kubenswrapper[4751]: I1002 12:15:02.119438 4751 generic.go:334] "Generic (PLEG): container finished" podID="d68a88b7-2660-43b8-8a32-512eeefb5e66" containerID="8638a7e82c89802b5cbedd24be5e686a5cd0a38f27f1ab2844ab8b304c8f5602" exitCode=0 Oct 02 12:15:02 crc kubenswrapper[4751]: I1002 12:15:02.119490 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" event={"ID":"d68a88b7-2660-43b8-8a32-512eeefb5e66","Type":"ContainerDied","Data":"8638a7e82c89802b5cbedd24be5e686a5cd0a38f27f1ab2844ab8b304c8f5602"} Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.859974 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.955990 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2jzb\" (UniqueName: \"kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb\") pod \"d68a88b7-2660-43b8-8a32-512eeefb5e66\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.956032 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume\") pod \"d68a88b7-2660-43b8-8a32-512eeefb5e66\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.956120 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume\") pod \"d68a88b7-2660-43b8-8a32-512eeefb5e66\" (UID: \"d68a88b7-2660-43b8-8a32-512eeefb5e66\") " Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.957917 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume" (OuterVolumeSpecName: "config-volume") pod "d68a88b7-2660-43b8-8a32-512eeefb5e66" (UID: "d68a88b7-2660-43b8-8a32-512eeefb5e66"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.962331 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb" (OuterVolumeSpecName: "kube-api-access-d2jzb") pod "d68a88b7-2660-43b8-8a32-512eeefb5e66" (UID: "d68a88b7-2660-43b8-8a32-512eeefb5e66"). InnerVolumeSpecName "kube-api-access-d2jzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:15:03 crc kubenswrapper[4751]: I1002 12:15:03.963315 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d68a88b7-2660-43b8-8a32-512eeefb5e66" (UID: "d68a88b7-2660-43b8-8a32-512eeefb5e66"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.058496 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2jzb\" (UniqueName: \"kubernetes.io/projected/d68a88b7-2660-43b8-8a32-512eeefb5e66-kube-api-access-d2jzb\") on node \"crc\" DevicePath \"\"" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.058579 4751 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d68a88b7-2660-43b8-8a32-512eeefb5e66-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.058592 4751 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d68a88b7-2660-43b8-8a32-512eeefb5e66-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.141759 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" event={"ID":"d68a88b7-2660-43b8-8a32-512eeefb5e66","Type":"ContainerDied","Data":"bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73"} Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.141810 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9fae18cf720929e8461bc1871ed7817d02e8b5c079637042e833c6d5655e73" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.141823 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323455-dkd7t" Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.953599 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9"] Oct 02 12:15:04 crc kubenswrapper[4751]: I1002 12:15:04.960545 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323410-5zjb9"] Oct 02 12:15:05 crc kubenswrapper[4751]: I1002 12:15:05.572331 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7456eeb2-be8e-4c1e-a175-ea64db231d46" path="/var/lib/kubelet/pods/7456eeb2-be8e-4c1e-a175-ea64db231d46/volumes" Oct 02 12:15:12 crc kubenswrapper[4751]: I1002 12:15:12.823240 4751 scope.go:117] "RemoveContainer" containerID="4890ecf1811e2780779ec180940eb2c29b71e6751777e974ef96dafb26cb0d89" Oct 02 12:15:31 crc kubenswrapper[4751]: I1002 12:15:31.507612 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:15:31 crc kubenswrapper[4751]: I1002 12:15:31.508313 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:16:01 crc kubenswrapper[4751]: I1002 12:16:01.506941 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:16:01 crc kubenswrapper[4751]: I1002 12:16:01.507583 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:16:31 crc kubenswrapper[4751]: I1002 12:16:31.508623 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:16:31 crc kubenswrapper[4751]: I1002 12:16:31.509284 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:16:31 crc kubenswrapper[4751]: I1002 12:16:31.509346 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:16:31 crc kubenswrapper[4751]: I1002 12:16:31.511012 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:16:31 crc kubenswrapper[4751]: I1002 12:16:31.511112 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" gracePeriod=600 Oct 02 12:16:31 crc kubenswrapper[4751]: E1002 12:16:31.675299 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:16:32 crc kubenswrapper[4751]: I1002 12:16:32.116898 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" exitCode=0 Oct 02 12:16:32 crc kubenswrapper[4751]: I1002 12:16:32.116992 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4"} Oct 02 12:16:32 crc kubenswrapper[4751]: I1002 12:16:32.117433 4751 scope.go:117] "RemoveContainer" containerID="0e7b1f7bfbefaf1d64d2d99db29820461d9779d68b50e9cd9e2ec9aa1187850d" Oct 02 12:16:32 crc kubenswrapper[4751]: I1002 12:16:32.118149 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:16:32 crc kubenswrapper[4751]: E1002 12:16:32.118588 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:16:42 crc kubenswrapper[4751]: I1002 12:16:42.551227 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:16:42 crc kubenswrapper[4751]: E1002 12:16:42.552163 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:16:56 crc kubenswrapper[4751]: I1002 12:16:56.551101 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:16:56 crc kubenswrapper[4751]: E1002 12:16:56.553921 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:17:08 crc kubenswrapper[4751]: I1002 12:17:08.551048 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:17:08 crc kubenswrapper[4751]: E1002 12:17:08.552231 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:17:20 crc kubenswrapper[4751]: I1002 12:17:20.551799 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:17:20 crc kubenswrapper[4751]: E1002 12:17:20.553497 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:17:34 crc kubenswrapper[4751]: I1002 12:17:34.551140 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:17:34 crc kubenswrapper[4751]: E1002 12:17:34.552211 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:17:48 crc kubenswrapper[4751]: I1002 12:17:48.550204 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:17:48 crc kubenswrapper[4751]: E1002 12:17:48.550982 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:18:03 crc kubenswrapper[4751]: I1002 12:18:03.552459 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:18:03 crc kubenswrapper[4751]: E1002 12:18:03.553484 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:18:18 crc kubenswrapper[4751]: I1002 12:18:18.550787 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:18:18 crc kubenswrapper[4751]: E1002 12:18:18.551902 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:18:30 crc kubenswrapper[4751]: I1002 12:18:30.550477 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:18:30 crc kubenswrapper[4751]: E1002 12:18:30.551281 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:18:41 crc kubenswrapper[4751]: I1002 12:18:41.550899 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:18:41 crc kubenswrapper[4751]: E1002 12:18:41.552938 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:18:56 crc kubenswrapper[4751]: I1002 12:18:56.552122 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:18:56 crc kubenswrapper[4751]: E1002 12:18:56.553000 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:19:09 crc kubenswrapper[4751]: I1002 12:19:09.557960 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:19:09 crc kubenswrapper[4751]: E1002 12:19:09.558940 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:19:22 crc kubenswrapper[4751]: I1002 12:19:22.550261 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:19:22 crc kubenswrapper[4751]: E1002 12:19:22.551371 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:19:36 crc kubenswrapper[4751]: I1002 12:19:36.550389 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:19:36 crc kubenswrapper[4751]: E1002 12:19:36.551310 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:19:50 crc kubenswrapper[4751]: I1002 12:19:50.550836 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:19:50 crc kubenswrapper[4751]: E1002 12:19:50.551710 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:03 crc kubenswrapper[4751]: I1002 12:20:03.550693 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:20:03 crc kubenswrapper[4751]: E1002 12:20:03.551496 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:17 crc kubenswrapper[4751]: I1002 12:20:17.550769 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:20:17 crc kubenswrapper[4751]: E1002 12:20:17.551694 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:28 crc kubenswrapper[4751]: I1002 12:20:28.550937 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:20:28 crc kubenswrapper[4751]: E1002 12:20:28.552054 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.527299 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5wn2f/must-gather-dqrtt"] Oct 02 12:20:34 crc kubenswrapper[4751]: E1002 12:20:34.528230 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68a88b7-2660-43b8-8a32-512eeefb5e66" containerName="collect-profiles" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.528243 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68a88b7-2660-43b8-8a32-512eeefb5e66" containerName="collect-profiles" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.528451 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68a88b7-2660-43b8-8a32-512eeefb5e66" containerName="collect-profiles" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.529692 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.535038 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5wn2f"/"openshift-service-ca.crt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.535235 4751 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5wn2f"/"default-dockercfg-n9mck" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.535544 4751 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5wn2f"/"kube-root-ca.crt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.539746 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5wn2f/must-gather-dqrtt"] Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.571537 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqvs4\" (UniqueName: \"kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.572440 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.673757 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqvs4\" (UniqueName: \"kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.674138 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.674715 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.696891 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqvs4\" (UniqueName: \"kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4\") pod \"must-gather-dqrtt\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:34 crc kubenswrapper[4751]: I1002 12:20:34.854868 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:20:35 crc kubenswrapper[4751]: I1002 12:20:35.338605 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5wn2f/must-gather-dqrtt"] Oct 02 12:20:35 crc kubenswrapper[4751]: I1002 12:20:35.346698 4751 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 12:20:35 crc kubenswrapper[4751]: I1002 12:20:35.667497 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" event={"ID":"2dc32e3a-c701-45f4-bdad-74dab7cfa711","Type":"ContainerStarted","Data":"c18561a370dc7eec25e7f1cdce989181f0b2f0ebdd01a4de177772015748600c"} Oct 02 12:20:36 crc kubenswrapper[4751]: I1002 12:20:36.905318 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:20:36 crc kubenswrapper[4751]: I1002 12:20:36.907898 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:36 crc kubenswrapper[4751]: I1002 12:20:36.913337 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.017134 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fmwf\" (UniqueName: \"kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.017274 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.017330 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.118925 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fmwf\" (UniqueName: \"kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.119007 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.119045 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.119546 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.119572 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.142152 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fmwf\" (UniqueName: \"kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf\") pod \"certified-operators-hnr8r\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:37 crc kubenswrapper[4751]: I1002 12:20:37.248009 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:38 crc kubenswrapper[4751]: I1002 12:20:38.588192 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:20:42 crc kubenswrapper[4751]: I1002 12:20:42.549792 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:20:42 crc kubenswrapper[4751]: E1002 12:20:42.551654 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:43 crc kubenswrapper[4751]: I1002 12:20:43.753022 4751 generic.go:334] "Generic (PLEG): container finished" podID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerID="d54b082a9ea5a157cfdbfcc7541c1ccb29ae9eb095ae1d6f90ca399acbc369a7" exitCode=0 Oct 02 12:20:43 crc kubenswrapper[4751]: I1002 12:20:43.753275 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerDied","Data":"d54b082a9ea5a157cfdbfcc7541c1ccb29ae9eb095ae1d6f90ca399acbc369a7"} Oct 02 12:20:43 crc kubenswrapper[4751]: I1002 12:20:43.753464 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerStarted","Data":"2037df163828f012761b3a0347505f69e9016a0d4fb7f26b2169e62969a851ce"} Oct 02 12:20:44 crc kubenswrapper[4751]: I1002 12:20:44.763924 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" event={"ID":"2dc32e3a-c701-45f4-bdad-74dab7cfa711","Type":"ContainerStarted","Data":"aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8"} Oct 02 12:20:44 crc kubenswrapper[4751]: I1002 12:20:44.764329 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" event={"ID":"2dc32e3a-c701-45f4-bdad-74dab7cfa711","Type":"ContainerStarted","Data":"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92"} Oct 02 12:20:44 crc kubenswrapper[4751]: I1002 12:20:44.794942 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" podStartSLOduration=1.943570841 podStartE2EDuration="10.794918801s" podCreationTimestamp="2025-10-02 12:20:34 +0000 UTC" firstStartedPulling="2025-10-02 12:20:35.34642621 +0000 UTC m=+5317.400652670" lastFinishedPulling="2025-10-02 12:20:44.19777418 +0000 UTC m=+5326.252000630" observedRunningTime="2025-10-02 12:20:44.780212876 +0000 UTC m=+5326.834439326" watchObservedRunningTime="2025-10-02 12:20:44.794918801 +0000 UTC m=+5326.849145261" Oct 02 12:20:46 crc kubenswrapper[4751]: I1002 12:20:46.784637 4751 generic.go:334] "Generic (PLEG): container finished" podID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerID="eea132963f1088755b5e8f2a9a6f3b675eee4f688c933a1f5977470960a8e1e6" exitCode=0 Oct 02 12:20:46 crc kubenswrapper[4751]: I1002 12:20:46.784719 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerDied","Data":"eea132963f1088755b5e8f2a9a6f3b675eee4f688c933a1f5977470960a8e1e6"} Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.288661 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-n82g8"] Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.291306 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.442994 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh4qb\" (UniqueName: \"kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.443550 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.545410 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh4qb\" (UniqueName: \"kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.545880 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.545974 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.577563 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh4qb\" (UniqueName: \"kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb\") pod \"crc-debug-n82g8\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.615037 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:20:48 crc kubenswrapper[4751]: W1002 12:20:48.677126 4751 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedcbd243_0b0e_4de2_8c7f_55b19ff06b8d.slice/crio-008dce531dff771bef1ef3bcdb4e7d1000566aa4b4f89ece7e46d19d65b742b6 WatchSource:0}: Error finding container 008dce531dff771bef1ef3bcdb4e7d1000566aa4b4f89ece7e46d19d65b742b6: Status 404 returned error can't find the container with id 008dce531dff771bef1ef3bcdb4e7d1000566aa4b4f89ece7e46d19d65b742b6 Oct 02 12:20:48 crc kubenswrapper[4751]: I1002 12:20:48.805048 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" event={"ID":"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d","Type":"ContainerStarted","Data":"008dce531dff771bef1ef3bcdb4e7d1000566aa4b4f89ece7e46d19d65b742b6"} Oct 02 12:20:49 crc kubenswrapper[4751]: I1002 12:20:49.820685 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerStarted","Data":"cd8e0c7f3c56cc48b152a71ea571312aefda758ede2149c7a846a81282b7b4b1"} Oct 02 12:20:49 crc kubenswrapper[4751]: I1002 12:20:49.839129 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hnr8r" podStartSLOduration=9.953534703999999 podStartE2EDuration="13.839113242s" podCreationTimestamp="2025-10-02 12:20:36 +0000 UTC" firstStartedPulling="2025-10-02 12:20:44.765810958 +0000 UTC m=+5326.820037408" lastFinishedPulling="2025-10-02 12:20:48.651389496 +0000 UTC m=+5330.705615946" observedRunningTime="2025-10-02 12:20:49.837212721 +0000 UTC m=+5331.891439181" watchObservedRunningTime="2025-10-02 12:20:49.839113242 +0000 UTC m=+5331.893339692" Oct 02 12:20:56 crc kubenswrapper[4751]: I1002 12:20:56.550376 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:20:56 crc kubenswrapper[4751]: E1002 12:20:56.551027 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:20:57 crc kubenswrapper[4751]: I1002 12:20:57.249183 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:57 crc kubenswrapper[4751]: I1002 12:20:57.249483 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:20:58 crc kubenswrapper[4751]: I1002 12:20:58.301145 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hnr8r" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" probeResult="failure" output=< Oct 02 12:20:58 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 12:20:58 crc kubenswrapper[4751]: > Oct 02 12:21:03 crc kubenswrapper[4751]: E1002 12:21:03.969494 4751 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Oct 02 12:21:03 crc kubenswrapper[4751]: E1002 12:21:03.970144 4751 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jh4qb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-n82g8_openshift-must-gather-5wn2f(edcbd243-0b0e-4de2-8c7f-55b19ff06b8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 12:21:03 crc kubenswrapper[4751]: E1002 12:21:03.971351 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" Oct 02 12:21:04 crc kubenswrapper[4751]: E1002 12:21:04.959796 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" Oct 02 12:21:08 crc kubenswrapper[4751]: I1002 12:21:08.310155 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hnr8r" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" probeResult="failure" output=< Oct 02 12:21:08 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 12:21:08 crc kubenswrapper[4751]: > Oct 02 12:21:09 crc kubenswrapper[4751]: I1002 12:21:09.560931 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:21:09 crc kubenswrapper[4751]: E1002 12:21:09.561538 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:21:18 crc kubenswrapper[4751]: I1002 12:21:18.298836 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hnr8r" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" probeResult="failure" output=< Oct 02 12:21:18 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 12:21:18 crc kubenswrapper[4751]: > Oct 02 12:21:21 crc kubenswrapper[4751]: I1002 12:21:21.119632 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" event={"ID":"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d","Type":"ContainerStarted","Data":"879ee356a9e49164903fb6d4d89d30ddf9ffba8a81c3485db87ca6cde6107252"} Oct 02 12:21:21 crc kubenswrapper[4751]: I1002 12:21:21.137117 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" podStartSLOduration=1.922326843 podStartE2EDuration="33.137093008s" podCreationTimestamp="2025-10-02 12:20:48 +0000 UTC" firstStartedPulling="2025-10-02 12:20:48.679184854 +0000 UTC m=+5330.733411304" lastFinishedPulling="2025-10-02 12:21:19.893951009 +0000 UTC m=+5361.948177469" observedRunningTime="2025-10-02 12:21:21.134393365 +0000 UTC m=+5363.188619825" watchObservedRunningTime="2025-10-02 12:21:21.137093008 +0000 UTC m=+5363.191319458" Oct 02 12:21:21 crc kubenswrapper[4751]: I1002 12:21:21.549815 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:21:21 crc kubenswrapper[4751]: E1002 12:21:21.550401 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:21:28 crc kubenswrapper[4751]: I1002 12:21:28.295838 4751 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hnr8r" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" probeResult="failure" output=< Oct 02 12:21:28 crc kubenswrapper[4751]: timeout: failed to connect service ":50051" within 1s Oct 02 12:21:28 crc kubenswrapper[4751]: > Oct 02 12:21:33 crc kubenswrapper[4751]: I1002 12:21:33.550452 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:21:34 crc kubenswrapper[4751]: I1002 12:21:34.232405 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec"} Oct 02 12:21:37 crc kubenswrapper[4751]: I1002 12:21:37.305369 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:21:37 crc kubenswrapper[4751]: I1002 12:21:37.380699 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:21:38 crc kubenswrapper[4751]: I1002 12:21:38.124023 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:21:39 crc kubenswrapper[4751]: I1002 12:21:39.278159 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hnr8r" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" containerID="cri-o://cd8e0c7f3c56cc48b152a71ea571312aefda758ede2149c7a846a81282b7b4b1" gracePeriod=2 Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.289558 4751 generic.go:334] "Generic (PLEG): container finished" podID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerID="cd8e0c7f3c56cc48b152a71ea571312aefda758ede2149c7a846a81282b7b4b1" exitCode=0 Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.289638 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerDied","Data":"cd8e0c7f3c56cc48b152a71ea571312aefda758ede2149c7a846a81282b7b4b1"} Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.290128 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnr8r" event={"ID":"228c2bd3-4dda-4efa-92d2-9358a539ed23","Type":"ContainerDied","Data":"2037df163828f012761b3a0347505f69e9016a0d4fb7f26b2169e62969a851ce"} Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.290194 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2037df163828f012761b3a0347505f69e9016a0d4fb7f26b2169e62969a851ce" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.481062 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.548157 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content\") pod \"228c2bd3-4dda-4efa-92d2-9358a539ed23\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.548544 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities\") pod \"228c2bd3-4dda-4efa-92d2-9358a539ed23\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.548599 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fmwf\" (UniqueName: \"kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf\") pod \"228c2bd3-4dda-4efa-92d2-9358a539ed23\" (UID: \"228c2bd3-4dda-4efa-92d2-9358a539ed23\") " Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.549124 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities" (OuterVolumeSpecName: "utilities") pod "228c2bd3-4dda-4efa-92d2-9358a539ed23" (UID: "228c2bd3-4dda-4efa-92d2-9358a539ed23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.555656 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf" (OuterVolumeSpecName: "kube-api-access-7fmwf") pod "228c2bd3-4dda-4efa-92d2-9358a539ed23" (UID: "228c2bd3-4dda-4efa-92d2-9358a539ed23"). InnerVolumeSpecName "kube-api-access-7fmwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.626479 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "228c2bd3-4dda-4efa-92d2-9358a539ed23" (UID: "228c2bd3-4dda-4efa-92d2-9358a539ed23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.653506 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.653535 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fmwf\" (UniqueName: \"kubernetes.io/projected/228c2bd3-4dda-4efa-92d2-9358a539ed23-kube-api-access-7fmwf\") on node \"crc\" DevicePath \"\"" Oct 02 12:21:40 crc kubenswrapper[4751]: I1002 12:21:40.653545 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/228c2bd3-4dda-4efa-92d2-9358a539ed23-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:21:41 crc kubenswrapper[4751]: I1002 12:21:41.297734 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnr8r" Oct 02 12:21:41 crc kubenswrapper[4751]: I1002 12:21:41.330697 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:21:41 crc kubenswrapper[4751]: I1002 12:21:41.338038 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hnr8r"] Oct 02 12:21:41 crc kubenswrapper[4751]: I1002 12:21:41.560334 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" path="/var/lib/kubelet/pods/228c2bd3-4dda-4efa-92d2-9358a539ed23/volumes" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.070857 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5565ddb6f8-hdw9n_a4115615-cab2-4426-ba9f-6bb96b1fae57/barbican-api/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.108118 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5565ddb6f8-hdw9n_a4115615-cab2-4426-ba9f-6bb96b1fae57/barbican-api-log/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.297514 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d89ffdd4b-n5d8c_c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38/barbican-keystone-listener/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.355294 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d89ffdd4b-n5d8c_c60aba27-b8d3-4c36-bd86-ac2a7d8e3c38/barbican-keystone-listener-log/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.506672 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c77b8fdb7-bhwjp_cb83ae02-2a40-4786-a2aa-159f7c5a9141/barbican-worker/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.547530 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c77b8fdb7-bhwjp_cb83ae02-2a40-4786-a2aa-159f7c5a9141/barbican-worker-log/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.687966 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-xwmp9_fe7b2e26-7acf-4950-b080-7f5c5e866491/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.894952 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9a0ea287-4160-468d-b606-efdbc47e9c50/ceilometer-central-agent/1.log" Oct 02 12:21:51 crc kubenswrapper[4751]: I1002 12:21:51.936893 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9a0ea287-4160-468d-b606-efdbc47e9c50/ceilometer-central-agent/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.024659 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9a0ea287-4160-468d-b606-efdbc47e9c50/ceilometer-notification-agent/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.098600 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9a0ea287-4160-468d-b606-efdbc47e9c50/proxy-httpd/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.128448 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9a0ea287-4160-468d-b606-efdbc47e9c50/sg-core/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.330375 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9389a471-3750-414f-889f-389fbb54e248/cinder-api-log/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.406654 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9389a471-3750-414f-889f-389fbb54e248/cinder-api/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.544574 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9f7323c9-33fe-466b-9544-85a75f09938d/cinder-scheduler/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.634803 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9f7323c9-33fe-466b-9544-85a75f09938d/probe/0.log" Oct 02 12:21:52 crc kubenswrapper[4751]: I1002 12:21:52.748519 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dpbh2_377f0b22-1de1-49d9-a4d6-f48bdad67770/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.215311 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7g2tk_0b5f4bef-7365-4e42-8eb4-2165c653e49a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.306756 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-b9hdc_6768867d-b798-4bb1-be1d-d329d01b5a7f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.401242 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-slxrn_af3ed200-3eef-4cc4-8b65-a0b92d3a8b45/init/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.626957 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-slxrn_af3ed200-3eef-4cc4-8b65-a0b92d3a8b45/init/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.676943 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-slxrn_af3ed200-3eef-4cc4-8b65-a0b92d3a8b45/dnsmasq-dns/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.889805 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-hpq2l_ffb43401-318b-43e5-b381-01d9bbf445cc/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:53 crc kubenswrapper[4751]: I1002 12:21:53.947962 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f1a17d69-9670-4f40-89dd-01099655729a/glance-httpd/0.log" Oct 02 12:21:54 crc kubenswrapper[4751]: I1002 12:21:54.093515 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f1a17d69-9670-4f40-89dd-01099655729a/glance-log/0.log" Oct 02 12:21:54 crc kubenswrapper[4751]: I1002 12:21:54.258287 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4ded07b6-69f6-4fed-9993-5877f61a4bf8/glance-httpd/0.log" Oct 02 12:21:54 crc kubenswrapper[4751]: I1002 12:21:54.304517 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4ded07b6-69f6-4fed-9993-5877f61a4bf8/glance-log/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.081076 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8hpbk_3332ad0f-fbf3-455e-8cb9-892c3cb435c2/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.163769 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6cf9f465b8-rf9bv_3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd/horizon/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.347901 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-ntcj9_a3f64ee3-4f5a-4133-8c31-7f8a770623ee/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.574299 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323441-lp668_eb0b7201-88ed-4df9-ba6b-6f5e6b2e3f96/keystone-cron/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.653984 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7b49c7b4b6-cxfxp_f80ce64a-4124-4ce4-8ffe-66bdbaf3488b/keystone-api/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.812919 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_df30039f-e3d2-43ce-aab5-050e2881c224/kube-state-metrics/0.log" Oct 02 12:21:55 crc kubenswrapper[4751]: I1002 12:21:55.911212 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6cf9f465b8-rf9bv_3b70e752-8ff4-47eb-ba3f-ae41af3fb8dd/horizon-log/0.log" Oct 02 12:21:56 crc kubenswrapper[4751]: I1002 12:21:56.278582 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9kkz9_ef947945-9630-4475-84ea-0798e9707829/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:56 crc kubenswrapper[4751]: I1002 12:21:56.735119 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b677f5d65-d5g4x_5bcbea36-e54d-4ea2-8911-6060ca0edf2e/neutron-api/0.log" Oct 02 12:21:56 crc kubenswrapper[4751]: I1002 12:21:56.903845 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b677f5d65-d5g4x_5bcbea36-e54d-4ea2-8911-6060ca0edf2e/neutron-httpd/0.log" Oct 02 12:21:57 crc kubenswrapper[4751]: I1002 12:21:57.111005 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-97ljf_8cfe9274-53ad-433f-b478-17b0b5b45bff/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:57 crc kubenswrapper[4751]: I1002 12:21:57.658282 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_609f5b43-0fb5-4044-95a2-ff1f0ab991ea/nova-api-log/0.log" Oct 02 12:21:57 crc kubenswrapper[4751]: I1002 12:21:57.943961 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_aa5e8e06-7063-4bb7-85dc-fe696c75aa7e/nova-cell0-conductor-conductor/0.log" Oct 02 12:21:58 crc kubenswrapper[4751]: I1002 12:21:58.244688 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_609f5b43-0fb5-4044-95a2-ff1f0ab991ea/nova-api-api/0.log" Oct 02 12:21:58 crc kubenswrapper[4751]: I1002 12:21:58.357946 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_03553bb2-dd3a-4362-946d-4379d45afa2d/nova-cell1-conductor-conductor/0.log" Oct 02 12:21:58 crc kubenswrapper[4751]: I1002 12:21:58.617510 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-24tk4_43932132-4f36-453c-82c7-8d48ced7e943/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:58 crc kubenswrapper[4751]: I1002 12:21:58.623647 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9596583b-e70e-435f-b390-0f873a1da605/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 12:21:58 crc kubenswrapper[4751]: I1002 12:21:58.864306 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-6hxbd_2cfb9cdb-16e0-4e8b-bb3f-28cefa9daadd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:59 crc kubenswrapper[4751]: I1002 12:21:59.202060 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-bnp94_745cbea8-a96d-45d6-baf9-13c0ef35df34/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:59 crc kubenswrapper[4751]: I1002 12:21:59.211434 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-kxw7z_aeebd860-2b84-42c5-a4f4-2bd00d9a9de7/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:59 crc kubenswrapper[4751]: I1002 12:21:59.678004 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-pmzbx_f861cad6-acad-4877-8054-e19e17de22ad/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:21:59 crc kubenswrapper[4751]: I1002 12:21:59.845277 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-qtgmg_839f86ef-2cd3-4c28-9b41-edc3b7dbcbac/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:00 crc kubenswrapper[4751]: I1002 12:22:00.005735 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-z2s8v_ecb4700e-df83-4e6d-be73-132f8ca80b09/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:00 crc kubenswrapper[4751]: I1002 12:22:00.251022 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2/nova-metadata-log/0.log" Oct 02 12:22:00 crc kubenswrapper[4751]: I1002 12:22:00.674232 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3fb989ac-8357-4e3c-af91-5378e0a874a9/nova-scheduler-scheduler/0.log" Oct 02 12:22:00 crc kubenswrapper[4751]: I1002 12:22:00.890611 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fa2f2972-7b83-44e7-8ba5-37fb04475f5c/mysql-bootstrap/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.121664 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fa2f2972-7b83-44e7-8ba5-37fb04475f5c/galera/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.150914 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_fa2f2972-7b83-44e7-8ba5-37fb04475f5c/mysql-bootstrap/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.381090 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6bd070ec-aa1f-4991-a239-df38cfff5fcf/mysql-bootstrap/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.669079 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6bd070ec-aa1f-4991-a239-df38cfff5fcf/mysql-bootstrap/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.742442 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6bd070ec-aa1f-4991-a239-df38cfff5fcf/galera/0.log" Oct 02 12:22:01 crc kubenswrapper[4751]: I1002 12:22:01.964398 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4002bd9a-1957-499c-b077-3f1eee6b239e/openstackclient/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.213592 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jsrq2_a85a9e69-5573-4806-ab4b-0b10c57d9c91/openstack-network-exporter/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.413131 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1cd18cd0-71aa-44a1-a3d9-87eabb6d55f2/nova-metadata-metadata/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.466255 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t54r5_7a41c88c-74de-4518-8313-441084f2cf0f/ovsdb-server-init/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.700337 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t54r5_7a41c88c-74de-4518-8313-441084f2cf0f/ovsdb-server-init/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.748671 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t54r5_7a41c88c-74de-4518-8313-441084f2cf0f/ovsdb-server/0.log" Oct 02 12:22:02 crc kubenswrapper[4751]: I1002 12:22:02.750077 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t54r5_7a41c88c-74de-4518-8313-441084f2cf0f/ovs-vswitchd/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.152921 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tqqkd_868943fd-a908-4e94-bef5-737ba8ebc1da/ovn-controller/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.345869 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8txlq_bf9e4b5e-949a-4a8b-a8fe-dea52bd10141/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.411058 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53ba58f6-e55e-4bd6-9365-04fe987aec95/openstack-network-exporter/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.562796 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_53ba58f6-e55e-4bd6-9365-04fe987aec95/ovn-northd/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.718123 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_539d6784-c995-47a3-9a80-53223cd9da09/openstack-network-exporter/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.805632 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_539d6784-c995-47a3-9a80-53223cd9da09/ovsdbserver-nb/0.log" Oct 02 12:22:03 crc kubenswrapper[4751]: I1002 12:22:03.929456 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6b3aa529-c99f-464f-85a0-70711c59f08f/openstack-network-exporter/0.log" Oct 02 12:22:04 crc kubenswrapper[4751]: I1002 12:22:04.117970 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6b3aa529-c99f-464f-85a0-70711c59f08f/ovsdbserver-sb/0.log" Oct 02 12:22:04 crc kubenswrapper[4751]: I1002 12:22:04.817681 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d856cc8dd-n7jjw_c9200866-1ba8-48d2-b271-9b55482ea983/placement-log/0.log" Oct 02 12:22:04 crc kubenswrapper[4751]: I1002 12:22:04.927290 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d856cc8dd-n7jjw_c9200866-1ba8-48d2-b271-9b55482ea983/placement-api/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.096197 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d34d3499-dc03-46d9-8dde-43c0f06c659b/setup-container/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.279766 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d34d3499-dc03-46d9-8dde-43c0f06c659b/setup-container/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.376857 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d34d3499-dc03-46d9-8dde-43c0f06c659b/rabbitmq/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.551999 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_90870cce-dc95-4298-8220-a9e0e61f315a/setup-container/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.697999 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_90870cce-dc95-4298-8220-a9e0e61f315a/setup-container/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.738647 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_90870cce-dc95-4298-8220-a9e0e61f315a/rabbitmq/0.log" Oct 02 12:22:05 crc kubenswrapper[4751]: I1002 12:22:05.949120 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-v47bs_400bc737-3667-4245-a0a3-fa5d2935d967/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:06 crc kubenswrapper[4751]: I1002 12:22:06.094893 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mqgg6_bb40236f-9313-40db-9e49-ffe654fdd92e/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:06 crc kubenswrapper[4751]: I1002 12:22:06.314334 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5dmkb_7928fef2-025b-4bab-ab81-839abc70715b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:06 crc kubenswrapper[4751]: I1002 12:22:06.520037 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-fwn4q_8dbf0347-7940-48ca-ad8c-29329d4b0391/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:06 crc kubenswrapper[4751]: I1002 12:22:06.655649 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-fk6rt_a3bc6e94-6738-40cf-9d23-e14ab0edc63b/ssh-known-hosts-edpm-deployment/0.log" Oct 02 12:22:06 crc kubenswrapper[4751]: I1002 12:22:06.936293 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-67bfc4fc59-jsm55_bf4cac44-422b-4c94-9275-5f952d606198/proxy-server/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.092901 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-67bfc4fc59-jsm55_bf4cac44-422b-4c94-9275-5f952d606198/proxy-httpd/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.168373 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bkkpx_30a668e4-d395-4b78-a014-279fd35ed304/swift-ring-rebalance/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.319492 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/account-auditor/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.375773 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/account-reaper/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.538765 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/account-server/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.562835 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/account-replicator/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.585652 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/container-auditor/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.772304 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/container-server/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.781741 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/container-replicator/0.log" Oct 02 12:22:07 crc kubenswrapper[4751]: I1002 12:22:07.853285 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/container-updater/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.002757 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/object-auditor/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.008709 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/object-expirer/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.069026 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/object-replicator/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.230397 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/object-updater/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.242671 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/object-server/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.315775 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/rsync/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.455399 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_b0071ad0-2497-45b8-bddc-9164ee537d18/swift-recon-cron/0.log" Oct 02 12:22:08 crc kubenswrapper[4751]: I1002 12:22:08.574191 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-dmz5f_74695503-3525-42a8-ad11-0e6644eebcee/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 12:22:11 crc kubenswrapper[4751]: I1002 12:22:11.478674 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2b5f725c-7545-484d-a3c4-37e69cf21a89/memcached/0.log" Oct 02 12:23:06 crc kubenswrapper[4751]: I1002 12:23:06.236006 4751 generic.go:334] "Generic (PLEG): container finished" podID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" containerID="879ee356a9e49164903fb6d4d89d30ddf9ffba8a81c3485db87ca6cde6107252" exitCode=0 Oct 02 12:23:06 crc kubenswrapper[4751]: I1002 12:23:06.236071 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" event={"ID":"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d","Type":"ContainerDied","Data":"879ee356a9e49164903fb6d4d89d30ddf9ffba8a81c3485db87ca6cde6107252"} Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.389297 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.422110 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-n82g8"] Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.429349 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-n82g8"] Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.494100 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host\") pod \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.494196 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh4qb\" (UniqueName: \"kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb\") pod \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\" (UID: \"edcbd243-0b0e-4de2-8c7f-55b19ff06b8d\") " Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.494523 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host" (OuterVolumeSpecName: "host") pod "edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" (UID: "edcbd243-0b0e-4de2-8c7f-55b19ff06b8d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.494788 4751 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-host\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.503996 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb" (OuterVolumeSpecName: "kube-api-access-jh4qb") pod "edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" (UID: "edcbd243-0b0e-4de2-8c7f-55b19ff06b8d"). InnerVolumeSpecName "kube-api-access-jh4qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.562983 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" path="/var/lib/kubelet/pods/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d/volumes" Oct 02 12:23:07 crc kubenswrapper[4751]: I1002 12:23:07.597206 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh4qb\" (UniqueName: \"kubernetes.io/projected/edcbd243-0b0e-4de2-8c7f-55b19ff06b8d-kube-api-access-jh4qb\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.258393 4751 scope.go:117] "RemoveContainer" containerID="879ee356a9e49164903fb6d4d89d30ddf9ffba8a81c3485db87ca6cde6107252" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.258468 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-n82g8" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.647969 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-2xdwf"] Oct 02 12:23:08 crc kubenswrapper[4751]: E1002 12:23:08.648601 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" containerName="container-00" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.648624 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" containerName="container-00" Oct 02 12:23:08 crc kubenswrapper[4751]: E1002 12:23:08.648673 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="extract-content" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.648686 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="extract-content" Oct 02 12:23:08 crc kubenswrapper[4751]: E1002 12:23:08.648713 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="extract-utilities" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.648726 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="extract-utilities" Oct 02 12:23:08 crc kubenswrapper[4751]: E1002 12:23:08.648762 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.648779 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.649256 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="228c2bd3-4dda-4efa-92d2-9358a539ed23" containerName="registry-server" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.649289 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="edcbd243-0b0e-4de2-8c7f-55b19ff06b8d" containerName="container-00" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.650512 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.722594 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.722701 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wvmc\" (UniqueName: \"kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.824679 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wvmc\" (UniqueName: \"kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.824934 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.825143 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.844617 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wvmc\" (UniqueName: \"kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc\") pod \"crc-debug-2xdwf\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:08 crc kubenswrapper[4751]: I1002 12:23:08.972841 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:09 crc kubenswrapper[4751]: I1002 12:23:09.278823 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" event={"ID":"89d4e23a-f8df-474d-ba19-5f9d27b354f0","Type":"ContainerStarted","Data":"580e7d8f5bb14fb97a40cc41911a76b3e0d1c99d89005af0951104949abe0cfa"} Oct 02 12:23:10 crc kubenswrapper[4751]: I1002 12:23:10.289489 4751 generic.go:334] "Generic (PLEG): container finished" podID="89d4e23a-f8df-474d-ba19-5f9d27b354f0" containerID="10392d1c9384d7afbf528f14943101ccda82ea8d52486540ab3f82dada485777" exitCode=0 Oct 02 12:23:10 crc kubenswrapper[4751]: I1002 12:23:10.289576 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" event={"ID":"89d4e23a-f8df-474d-ba19-5f9d27b354f0","Type":"ContainerDied","Data":"10392d1c9384d7afbf528f14943101ccda82ea8d52486540ab3f82dada485777"} Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.398652 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.513861 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wvmc\" (UniqueName: \"kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc\") pod \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.514152 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host\") pod \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\" (UID: \"89d4e23a-f8df-474d-ba19-5f9d27b354f0\") " Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.515323 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host" (OuterVolumeSpecName: "host") pod "89d4e23a-f8df-474d-ba19-5f9d27b354f0" (UID: "89d4e23a-f8df-474d-ba19-5f9d27b354f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.531423 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc" (OuterVolumeSpecName: "kube-api-access-5wvmc") pod "89d4e23a-f8df-474d-ba19-5f9d27b354f0" (UID: "89d4e23a-f8df-474d-ba19-5f9d27b354f0"). InnerVolumeSpecName "kube-api-access-5wvmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.616646 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wvmc\" (UniqueName: \"kubernetes.io/projected/89d4e23a-f8df-474d-ba19-5f9d27b354f0-kube-api-access-5wvmc\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:11 crc kubenswrapper[4751]: I1002 12:23:11.616987 4751 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/89d4e23a-f8df-474d-ba19-5f9d27b354f0-host\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:12 crc kubenswrapper[4751]: I1002 12:23:12.311148 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" event={"ID":"89d4e23a-f8df-474d-ba19-5f9d27b354f0","Type":"ContainerDied","Data":"580e7d8f5bb14fb97a40cc41911a76b3e0d1c99d89005af0951104949abe0cfa"} Oct 02 12:23:12 crc kubenswrapper[4751]: I1002 12:23:12.311213 4751 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580e7d8f5bb14fb97a40cc41911a76b3e0d1c99d89005af0951104949abe0cfa" Oct 02 12:23:12 crc kubenswrapper[4751]: I1002 12:23:12.311268 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-2xdwf" Oct 02 12:23:17 crc kubenswrapper[4751]: I1002 12:23:17.014805 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-2xdwf"] Oct 02 12:23:17 crc kubenswrapper[4751]: I1002 12:23:17.022917 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-2xdwf"] Oct 02 12:23:17 crc kubenswrapper[4751]: I1002 12:23:17.561641 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d4e23a-f8df-474d-ba19-5f9d27b354f0" path="/var/lib/kubelet/pods/89d4e23a-f8df-474d-ba19-5f9d27b354f0/volumes" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.197904 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-mp5hs"] Oct 02 12:23:18 crc kubenswrapper[4751]: E1002 12:23:18.199979 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d4e23a-f8df-474d-ba19-5f9d27b354f0" containerName="container-00" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.200088 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d4e23a-f8df-474d-ba19-5f9d27b354f0" containerName="container-00" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.200464 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d4e23a-f8df-474d-ba19-5f9d27b354f0" containerName="container-00" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.201267 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.332044 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmjn2\" (UniqueName: \"kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.332089 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.434295 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmjn2\" (UniqueName: \"kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.434338 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.434432 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.460384 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmjn2\" (UniqueName: \"kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2\") pod \"crc-debug-mp5hs\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:18 crc kubenswrapper[4751]: I1002 12:23:18.520747 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:19 crc kubenswrapper[4751]: I1002 12:23:19.370482 4751 generic.go:334] "Generic (PLEG): container finished" podID="26a32821-61af-4e9f-ba8d-f8764755b5e9" containerID="b73f6b0f15af490e5151c07c61fe31efb527b7ef4a05138e635701405d9cddcf" exitCode=0 Oct 02 12:23:19 crc kubenswrapper[4751]: I1002 12:23:19.370842 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" event={"ID":"26a32821-61af-4e9f-ba8d-f8764755b5e9","Type":"ContainerDied","Data":"b73f6b0f15af490e5151c07c61fe31efb527b7ef4a05138e635701405d9cddcf"} Oct 02 12:23:19 crc kubenswrapper[4751]: I1002 12:23:19.370881 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" event={"ID":"26a32821-61af-4e9f-ba8d-f8764755b5e9","Type":"ContainerStarted","Data":"56ead61c0d55dce3c0c3e5e8242d4f033cdb65a34bba604f29e0d5709eb8e63d"} Oct 02 12:23:19 crc kubenswrapper[4751]: I1002 12:23:19.413470 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-mp5hs"] Oct 02 12:23:19 crc kubenswrapper[4751]: I1002 12:23:19.421690 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5wn2f/crc-debug-mp5hs"] Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.493553 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.576156 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host\") pod \"26a32821-61af-4e9f-ba8d-f8764755b5e9\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.576281 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host" (OuterVolumeSpecName: "host") pod "26a32821-61af-4e9f-ba8d-f8764755b5e9" (UID: "26a32821-61af-4e9f-ba8d-f8764755b5e9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.576395 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmjn2\" (UniqueName: \"kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2\") pod \"26a32821-61af-4e9f-ba8d-f8764755b5e9\" (UID: \"26a32821-61af-4e9f-ba8d-f8764755b5e9\") " Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.576906 4751 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26a32821-61af-4e9f-ba8d-f8764755b5e9-host\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.586356 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2" (OuterVolumeSpecName: "kube-api-access-xmjn2") pod "26a32821-61af-4e9f-ba8d-f8764755b5e9" (UID: "26a32821-61af-4e9f-ba8d-f8764755b5e9"). InnerVolumeSpecName "kube-api-access-xmjn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:23:20 crc kubenswrapper[4751]: I1002 12:23:20.678747 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmjn2\" (UniqueName: \"kubernetes.io/projected/26a32821-61af-4e9f-ba8d-f8764755b5e9-kube-api-access-xmjn2\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.172804 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/util/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.296919 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/pull/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.309917 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/util/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.346148 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/pull/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.389761 4751 scope.go:117] "RemoveContainer" containerID="b73f6b0f15af490e5151c07c61fe31efb527b7ef4a05138e635701405d9cddcf" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.389794 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/crc-debug-mp5hs" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.509855 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/pull/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.559911 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a32821-61af-4e9f-ba8d-f8764755b5e9" path="/var/lib/kubelet/pods/26a32821-61af-4e9f-ba8d-f8764755b5e9/volumes" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.572940 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/extract/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.581457 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_157dbb73256b1b6d15c92cb3b9832917051f27d0aa325f8cd46370e26dbtrkk_b22b717b-33f6-403e-9541-2204576e92ae/util/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.701988 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-pm99x_ae064529-2f0b-4be8-93c0-9ac3cd3ab184/kube-rbac-proxy/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.796286 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kglk4_3bdcf078-737b-4a4e-9f17-10a55190af5b/kube-rbac-proxy/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.797472 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-pm99x_ae064529-2f0b-4be8-93c0-9ac3cd3ab184/manager/0.log" Oct 02 12:23:21 crc kubenswrapper[4751]: I1002 12:23:21.942594 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kglk4_3bdcf078-737b-4a4e-9f17-10a55190af5b/manager/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.001489 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rnd8b_e6b172bd-40be-498b-beae-119a3638c8e3/manager/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.015087 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rnd8b_e6b172bd-40be-498b-beae-119a3638c8e3/kube-rbac-proxy/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.075722 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:22 crc kubenswrapper[4751]: E1002 12:23:22.076185 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a32821-61af-4e9f-ba8d-f8764755b5e9" containerName="container-00" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.076200 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a32821-61af-4e9f-ba8d-f8764755b5e9" containerName="container-00" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.076415 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a32821-61af-4e9f-ba8d-f8764755b5e9" containerName="container-00" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.077725 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.087189 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.172744 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-clwc9_a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a/kube-rbac-proxy/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.206745 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94w45\" (UniqueName: \"kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.206945 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.207015 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.285719 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-clwc9_a40f4dc7-dd9c-40d2-a7d4-e2ad8dfe857a/manager/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.309437 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.309517 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94w45\" (UniqueName: \"kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.309660 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.310256 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.310303 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.335422 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94w45\" (UniqueName: \"kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45\") pod \"redhat-marketplace-b5clj\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.405292 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.441020 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-gcqh5_546a3f39-e64b-4747-806c-80b826920e8a/kube-rbac-proxy/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.578273 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-gcqh5_546a3f39-e64b-4747-806c-80b826920e8a/manager/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.697595 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-8jkm7_bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0/kube-rbac-proxy/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.699623 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.800356 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-8jkm7_bb6895fa-09f4-4a4c-9d06-0f0e60d97dd0/manager/0.log" Oct 02 12:23:22 crc kubenswrapper[4751]: I1002 12:23:22.804331 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-gqqxn_f512a98c-7304-4899-8382-65cab60c17a3/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.089256 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-gqqxn_f512a98c-7304-4899-8382-65cab60c17a3/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.105351 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-4rhhf_c6649ed7-0b11-4c16-8f0f-65cd3baa462e/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.113481 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-4rhhf_c6649ed7-0b11-4c16-8f0f-65cd3baa462e/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.279658 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-6lcbb_aec887c3-910f-4278-8834-afcc4c9b6aa3/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.379158 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-6lcbb_aec887c3-910f-4278-8834-afcc4c9b6aa3/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.410597 4751 generic.go:334] "Generic (PLEG): container finished" podID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerID="0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567" exitCode=0 Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.410654 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerDied","Data":"0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567"} Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.410681 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerStarted","Data":"d15d926b53112882425178aa5b29c888428d40870aef865fa04ab8c46daaf60d"} Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.495968 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-jk8xr_36f42236-589e-45ef-a69c-753d38fd54d6/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.513191 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-jk8xr_36f42236-589e-45ef-a69c-753d38fd54d6/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.612636 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-vq88r_2ed3e86e-2313-4efa-987a-d8209ce59d73/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.702925 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-vq88r_2ed3e86e-2313-4efa-987a-d8209ce59d73/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.758806 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-kqkzd_4905929c-80af-4de9-b7c7-4ba9f3b938ad/kube-rbac-proxy/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.861200 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-kqkzd_4905929c-80af-4de9-b7c7-4ba9f3b938ad/manager/0.log" Oct 02 12:23:23 crc kubenswrapper[4751]: I1002 12:23:23.917811 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hcnpc_f8b1c8b4-b491-44aa-885e-c6215289cb0d/kube-rbac-proxy/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.064736 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-hcnpc_f8b1c8b4-b491-44aa-885e-c6215289cb0d/manager/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.104727 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-vrndj_58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa/kube-rbac-proxy/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.111271 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-vrndj_58a05dc4-0f5e-43d9-bd9e-0a17ca619aaa/manager/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.284222 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-h7qc4_787aba9c-ec9d-4d98-bae0-77048675a98b/manager/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.325446 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-h7qc4_787aba9c-ec9d-4d98-bae0-77048675a98b/kube-rbac-proxy/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.417711 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f7d749dc7-j6vzk_5be43ec7-c21c-4a6c-a3ef-a816569fda32/kube-rbac-proxy/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.423917 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerStarted","Data":"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff"} Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.570393 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cc764bd77-ws526_fa36b949-67d9-4930-a31a-0381cd077c4a/kube-rbac-proxy/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.782220 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xpjtm_374515d5-bfd6-4830-bfd8-e5511e512187/registry-server/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.994757 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cc764bd77-ws526_fa36b949-67d9-4930-a31a-0381cd077c4a/operator/0.log" Oct 02 12:23:24 crc kubenswrapper[4751]: I1002 12:23:24.995567 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-flblb_64d6192e-f762-40ea-8d3f-9a2f8f27144b/kube-rbac-proxy/0.log" Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.121068 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-flblb_64d6192e-f762-40ea-8d3f-9a2f8f27144b/manager/0.log" Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.228401 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-j85jt_738b4521-cee5-4fa7-95df-3ac0ea68b26f/manager/0.log" Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.231699 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-j85jt_738b4521-cee5-4fa7-95df-3ac0ea68b26f/kube-rbac-proxy/0.log" Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.436114 4751 generic.go:334] "Generic (PLEG): container finished" podID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerID="9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff" exitCode=0 Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.436150 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerDied","Data":"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff"} Oct 02 12:23:25 crc kubenswrapper[4751]: I1002 12:23:25.885427 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-vz2v4_5a9de164-5d56-4ff5-ac4a-2cf16846bb4c/operator/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.027743 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-cqsln_91ad0328-73e0-4cf6-89d0-354d79055a1b/kube-rbac-proxy/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.042283 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-cqsln_91ad0328-73e0-4cf6-89d0-354d79055a1b/manager/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.138403 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f7d749dc7-j6vzk_5be43ec7-c21c-4a6c-a3ef-a816569fda32/manager/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.250206 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-pzqdp_42842b0c-bdf5-4b32-8c6d-f635c333c1b7/kube-rbac-proxy/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.278838 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-pzqdp_42842b0c-bdf5-4b32-8c6d-f635c333c1b7/manager/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.345802 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-q2bvx_77ffa158-937f-44d2-b65a-9f448a064288/kube-rbac-proxy/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.366279 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-q2bvx_77ffa158-937f-44d2-b65a-9f448a064288/manager/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.447741 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerStarted","Data":"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a"} Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.469695 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b5clj" podStartSLOduration=1.8394202210000001 podStartE2EDuration="4.469675524s" podCreationTimestamp="2025-10-02 12:23:22 +0000 UTC" firstStartedPulling="2025-10-02 12:23:23.412986979 +0000 UTC m=+5485.467213429" lastFinishedPulling="2025-10-02 12:23:26.043242282 +0000 UTC m=+5488.097468732" observedRunningTime="2025-10-02 12:23:26.465242324 +0000 UTC m=+5488.519468784" watchObservedRunningTime="2025-10-02 12:23:26.469675524 +0000 UTC m=+5488.523901974" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.499033 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-st8lx_a1b22cc9-2f48-475c-9325-fca7fbdfbc25/kube-rbac-proxy/0.log" Oct 02 12:23:26 crc kubenswrapper[4751]: I1002 12:23:26.517805 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-st8lx_a1b22cc9-2f48-475c-9325-fca7fbdfbc25/manager/0.log" Oct 02 12:23:32 crc kubenswrapper[4751]: I1002 12:23:32.406304 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:32 crc kubenswrapper[4751]: I1002 12:23:32.406860 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:32 crc kubenswrapper[4751]: I1002 12:23:32.462129 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:32 crc kubenswrapper[4751]: I1002 12:23:32.554028 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:32 crc kubenswrapper[4751]: I1002 12:23:32.695508 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:34 crc kubenswrapper[4751]: I1002 12:23:34.519892 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b5clj" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="registry-server" containerID="cri-o://0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a" gracePeriod=2 Oct 02 12:23:34 crc kubenswrapper[4751]: I1002 12:23:34.987112 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.153780 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities\") pod \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.154199 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content\") pod \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.154298 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94w45\" (UniqueName: \"kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45\") pod \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\" (UID: \"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b\") " Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.155024 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities" (OuterVolumeSpecName: "utilities") pod "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" (UID: "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.155588 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.161462 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45" (OuterVolumeSpecName: "kube-api-access-94w45") pod "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" (UID: "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b"). InnerVolumeSpecName "kube-api-access-94w45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.174258 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" (UID: "eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.257570 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.257823 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94w45\" (UniqueName: \"kubernetes.io/projected/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b-kube-api-access-94w45\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.532327 4751 generic.go:334] "Generic (PLEG): container finished" podID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerID="0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a" exitCode=0 Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.532398 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerDied","Data":"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a"} Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.532445 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5clj" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.533775 4751 scope.go:117] "RemoveContainer" containerID="0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.533680 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5clj" event={"ID":"eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b","Type":"ContainerDied","Data":"d15d926b53112882425178aa5b29c888428d40870aef865fa04ab8c46daaf60d"} Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.573215 4751 scope.go:117] "RemoveContainer" containerID="9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.585096 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.601636 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5clj"] Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.613931 4751 scope.go:117] "RemoveContainer" containerID="0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.666707 4751 scope.go:117] "RemoveContainer" containerID="0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a" Oct 02 12:23:35 crc kubenswrapper[4751]: E1002 12:23:35.667162 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a\": container with ID starting with 0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a not found: ID does not exist" containerID="0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.667223 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a"} err="failed to get container status \"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a\": rpc error: code = NotFound desc = could not find container \"0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a\": container with ID starting with 0bedabf66639c41c02353b66234571714bd3d86009eea11534d82c4015ba181a not found: ID does not exist" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.667248 4751 scope.go:117] "RemoveContainer" containerID="9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff" Oct 02 12:23:35 crc kubenswrapper[4751]: E1002 12:23:35.667512 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff\": container with ID starting with 9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff not found: ID does not exist" containerID="9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.667541 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff"} err="failed to get container status \"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff\": rpc error: code = NotFound desc = could not find container \"9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff\": container with ID starting with 9f7c1600ec41c8168c9d09697a26cc6f8451c3e93c3481aa9c80a1e01d0d41ff not found: ID does not exist" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.667556 4751 scope.go:117] "RemoveContainer" containerID="0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567" Oct 02 12:23:35 crc kubenswrapper[4751]: E1002 12:23:35.667844 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567\": container with ID starting with 0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567 not found: ID does not exist" containerID="0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567" Oct 02 12:23:35 crc kubenswrapper[4751]: I1002 12:23:35.667978 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567"} err="failed to get container status \"0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567\": rpc error: code = NotFound desc = could not find container \"0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567\": container with ID starting with 0595961b0ecfcee88e21f131222111f8bbaaa35305ef2edcb11e18b22b522567 not found: ID does not exist" Oct 02 12:23:37 crc kubenswrapper[4751]: I1002 12:23:37.559482 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" path="/var/lib/kubelet/pods/eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b/volumes" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.114814 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:41 crc kubenswrapper[4751]: E1002 12:23:41.116211 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="registry-server" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.116230 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="registry-server" Oct 02 12:23:41 crc kubenswrapper[4751]: E1002 12:23:41.116256 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="extract-content" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.116263 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="extract-content" Oct 02 12:23:41 crc kubenswrapper[4751]: E1002 12:23:41.116298 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="extract-utilities" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.116307 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="extract-utilities" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.116549 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafdaf39-c2c9-42c2-a80e-5d9ad9967b8b" containerName="registry-server" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.118569 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.127055 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.268310 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.268555 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.268602 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj7kk\" (UniqueName: \"kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.371047 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.371291 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.371344 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj7kk\" (UniqueName: \"kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.372381 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.372686 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.395914 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj7kk\" (UniqueName: \"kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk\") pod \"redhat-operators-mjnn9\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.442715 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.915032 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-p5jlr_0413ee62-e2ef-4ca0-a21f-2990f11169c3/control-plane-machine-set-operator/0.log" Oct 02 12:23:41 crc kubenswrapper[4751]: I1002 12:23:41.995273 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:42 crc kubenswrapper[4751]: I1002 12:23:42.207809 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4r69k_ed30f5ae-4c45-4928-9ef6-079587f62490/kube-rbac-proxy/0.log" Oct 02 12:23:42 crc kubenswrapper[4751]: I1002 12:23:42.226515 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4r69k_ed30f5ae-4c45-4928-9ef6-079587f62490/machine-api-operator/0.log" Oct 02 12:23:42 crc kubenswrapper[4751]: I1002 12:23:42.629149 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerID="b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3" exitCode=0 Oct 02 12:23:42 crc kubenswrapper[4751]: I1002 12:23:42.629297 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerDied","Data":"b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3"} Oct 02 12:23:42 crc kubenswrapper[4751]: I1002 12:23:42.629492 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerStarted","Data":"61e2fbd3f1e20a656890a8ee6cce09d50f9a4511301e804e26869a7c0fbea0f2"} Oct 02 12:23:44 crc kubenswrapper[4751]: I1002 12:23:44.675977 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerStarted","Data":"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759"} Oct 02 12:23:45 crc kubenswrapper[4751]: I1002 12:23:45.685978 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerID="9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759" exitCode=0 Oct 02 12:23:45 crc kubenswrapper[4751]: I1002 12:23:45.686524 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerDied","Data":"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759"} Oct 02 12:23:46 crc kubenswrapper[4751]: I1002 12:23:46.697518 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerStarted","Data":"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0"} Oct 02 12:23:46 crc kubenswrapper[4751]: I1002 12:23:46.724557 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mjnn9" podStartSLOduration=2.054653958 podStartE2EDuration="5.72453216s" podCreationTimestamp="2025-10-02 12:23:41 +0000 UTC" firstStartedPulling="2025-10-02 12:23:42.631417062 +0000 UTC m=+5504.685643512" lastFinishedPulling="2025-10-02 12:23:46.301295264 +0000 UTC m=+5508.355521714" observedRunningTime="2025-10-02 12:23:46.718696152 +0000 UTC m=+5508.772922602" watchObservedRunningTime="2025-10-02 12:23:46.72453216 +0000 UTC m=+5508.778758620" Oct 02 12:23:51 crc kubenswrapper[4751]: I1002 12:23:51.443599 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:51 crc kubenswrapper[4751]: I1002 12:23:51.444231 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:51 crc kubenswrapper[4751]: I1002 12:23:51.508032 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:51 crc kubenswrapper[4751]: I1002 12:23:51.799321 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:51 crc kubenswrapper[4751]: I1002 12:23:51.855209 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:53 crc kubenswrapper[4751]: I1002 12:23:53.672700 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-dn5sh_cef5a193-4c48-4da3-9518-0dd2fafc905a/cert-manager-controller/0.log" Oct 02 12:23:53 crc kubenswrapper[4751]: I1002 12:23:53.761393 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mjnn9" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="registry-server" containerID="cri-o://c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0" gracePeriod=2 Oct 02 12:23:53 crc kubenswrapper[4751]: I1002 12:23:53.882858 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cbzvj_b31d2144-af31-43e1-a625-2f558c8b21a9/cert-manager-cainjector/0.log" Oct 02 12:23:53 crc kubenswrapper[4751]: I1002 12:23:53.943796 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-m8wxt_aa39b082-35a9-4465-ae1f-8ef168d22ff4/cert-manager-webhook/0.log" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.232214 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.298445 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content\") pod \"8f9b0cde-2aba-46a6-916e-753f83f22f02\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.298596 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj7kk\" (UniqueName: \"kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk\") pod \"8f9b0cde-2aba-46a6-916e-753f83f22f02\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.298635 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities\") pod \"8f9b0cde-2aba-46a6-916e-753f83f22f02\" (UID: \"8f9b0cde-2aba-46a6-916e-753f83f22f02\") " Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.299337 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities" (OuterVolumeSpecName: "utilities") pod "8f9b0cde-2aba-46a6-916e-753f83f22f02" (UID: "8f9b0cde-2aba-46a6-916e-753f83f22f02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.306621 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk" (OuterVolumeSpecName: "kube-api-access-hj7kk") pod "8f9b0cde-2aba-46a6-916e-753f83f22f02" (UID: "8f9b0cde-2aba-46a6-916e-753f83f22f02"). InnerVolumeSpecName "kube-api-access-hj7kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.385113 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f9b0cde-2aba-46a6-916e-753f83f22f02" (UID: "8f9b0cde-2aba-46a6-916e-753f83f22f02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.401312 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.401351 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj7kk\" (UniqueName: \"kubernetes.io/projected/8f9b0cde-2aba-46a6-916e-753f83f22f02-kube-api-access-hj7kk\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.401362 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9b0cde-2aba-46a6-916e-753f83f22f02-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.781413 4751 generic.go:334] "Generic (PLEG): container finished" podID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerID="c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0" exitCode=0 Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.781733 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mjnn9" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.781643 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerDied","Data":"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0"} Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.782195 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mjnn9" event={"ID":"8f9b0cde-2aba-46a6-916e-753f83f22f02","Type":"ContainerDied","Data":"61e2fbd3f1e20a656890a8ee6cce09d50f9a4511301e804e26869a7c0fbea0f2"} Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.782237 4751 scope.go:117] "RemoveContainer" containerID="c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.830260 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.832034 4751 scope.go:117] "RemoveContainer" containerID="9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.836328 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mjnn9"] Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.856117 4751 scope.go:117] "RemoveContainer" containerID="b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.900495 4751 scope.go:117] "RemoveContainer" containerID="c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0" Oct 02 12:23:54 crc kubenswrapper[4751]: E1002 12:23:54.901069 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0\": container with ID starting with c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0 not found: ID does not exist" containerID="c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.901122 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0"} err="failed to get container status \"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0\": rpc error: code = NotFound desc = could not find container \"c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0\": container with ID starting with c09672f5797507295f1344440e34f50b9c8d91a932a61ce0c58e4bafa4b792c0 not found: ID does not exist" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.901157 4751 scope.go:117] "RemoveContainer" containerID="9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759" Oct 02 12:23:54 crc kubenswrapper[4751]: E1002 12:23:54.901607 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759\": container with ID starting with 9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759 not found: ID does not exist" containerID="9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.901639 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759"} err="failed to get container status \"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759\": rpc error: code = NotFound desc = could not find container \"9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759\": container with ID starting with 9fb55e46d4daf8e34409b9be6804a1f198aa0ff3ebf4722be92885e1e325d759 not found: ID does not exist" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.901658 4751 scope.go:117] "RemoveContainer" containerID="b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3" Oct 02 12:23:54 crc kubenswrapper[4751]: E1002 12:23:54.901978 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3\": container with ID starting with b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3 not found: ID does not exist" containerID="b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3" Oct 02 12:23:54 crc kubenswrapper[4751]: I1002 12:23:54.902008 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3"} err="failed to get container status \"b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3\": rpc error: code = NotFound desc = could not find container \"b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3\": container with ID starting with b63a00a7aba929930026e8c76a7dd244621b0bb9aca71c9ee0f1752777ff69e3 not found: ID does not exist" Oct 02 12:23:55 crc kubenswrapper[4751]: I1002 12:23:55.562982 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" path="/var/lib/kubelet/pods/8f9b0cde-2aba-46a6-916e-753f83f22f02/volumes" Oct 02 12:24:01 crc kubenswrapper[4751]: I1002 12:24:01.507237 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:24:01 crc kubenswrapper[4751]: I1002 12:24:01.507743 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.124952 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-dpwvm_fbc72901-fb4e-4476-b339-41c851edf1fe/nmstate-console-plugin/0.log" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.295639 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-czmv9_203e7187-9944-44f7-8a43-615567c3ab3c/nmstate-handler/0.log" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.315840 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6jmww_9b91b188-db18-4362-bf3e-612233c1aa7d/kube-rbac-proxy/0.log" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.400258 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6jmww_9b91b188-db18-4362-bf3e-612233c1aa7d/nmstate-metrics/0.log" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.487468 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-p576x_bf80f016-df98-4666-983e-d8537054d39d/nmstate-operator/0.log" Oct 02 12:24:05 crc kubenswrapper[4751]: I1002 12:24:05.601478 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-9tq9l_b0a29038-b5ae-46b0-b5a2-4f7f952fec45/nmstate-webhook/0.log" Oct 02 12:24:19 crc kubenswrapper[4751]: I1002 12:24:19.936524 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2dzwd_82465ad7-d9ca-4475-a0dc-6ce4068f20d2/kube-rbac-proxy/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.031896 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2dzwd_82465ad7-d9ca-4475-a0dc-6ce4068f20d2/controller/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.075142 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-frr-files/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.258189 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-reloader/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.299329 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-reloader/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.318130 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-frr-files/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.326511 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-metrics/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.456749 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-reloader/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.463681 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-frr-files/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.529538 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-metrics/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.538294 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-metrics/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.705352 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-frr-files/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.731548 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-reloader/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.767208 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/controller/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.772532 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/cp-metrics/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.934258 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/kube-rbac-proxy-frr/0.log" Oct 02 12:24:20 crc kubenswrapper[4751]: I1002 12:24:20.957506 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/frr-metrics/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.003020 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/kube-rbac-proxy/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.143736 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/reloader/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.273697 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pvfc5_6005ba7f-c783-4bc6-8df6-d045f6a9d80b/frr-k8s-webhook-server/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.387343 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66b7df5dd4-ghn54_4b846a35-2b40-4cfd-894c-94d9a3bd1a8e/manager/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.727474 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6c5dfc98c6-msmjp_4bdf3e34-7618-4ee2-bfc6-3fc09cbbdece/webhook-server/0.log" Oct 02 12:24:21 crc kubenswrapper[4751]: I1002 12:24:21.852706 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tpkx9_46040bd9-f688-4966-a105-8ad5f93a1c8a/kube-rbac-proxy/0.log" Oct 02 12:24:22 crc kubenswrapper[4751]: I1002 12:24:22.328068 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-cq9wn_c82c3ef0-d1d6-45d7-902b-100cfd482853/frr/0.log" Oct 02 12:24:22 crc kubenswrapper[4751]: I1002 12:24:22.397082 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tpkx9_46040bd9-f688-4966-a105-8ad5f93a1c8a/speaker/0.log" Oct 02 12:24:31 crc kubenswrapper[4751]: I1002 12:24:31.506862 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:24:31 crc kubenswrapper[4751]: I1002 12:24:31.507606 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:24:33 crc kubenswrapper[4751]: I1002 12:24:33.840315 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/util/0.log" Oct 02 12:24:33 crc kubenswrapper[4751]: I1002 12:24:33.952749 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/util/0.log" Oct 02 12:24:33 crc kubenswrapper[4751]: I1002 12:24:33.985330 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/pull/0.log" Oct 02 12:24:33 crc kubenswrapper[4751]: I1002 12:24:33.985659 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/pull/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.170186 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/extract/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.218752 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/util/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.220685 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2n6vvz_020abe52-0534-45cd-845e-5680d8359e6b/pull/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.331940 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-utilities/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.518572 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-content/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.522815 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-content/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.538227 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-utilities/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.711381 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-utilities/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.759880 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/extract-content/0.log" Oct 02 12:24:34 crc kubenswrapper[4751]: I1002 12:24:34.979616 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-utilities/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.315920 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-utilities/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.320727 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-content/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.322142 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-content/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.381141 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhk57_f17f0af2-1997-416f-a908-4653c87513a8/registry-server/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.500218 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-utilities/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.515746 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/extract-content/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.734306 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/util/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.907990 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/util/0.log" Oct 02 12:24:35 crc kubenswrapper[4751]: I1002 12:24:35.989066 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/pull/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.009265 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/pull/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.177495 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/util/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.225386 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/extract/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.287576 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz78b7_5cdb897a-9c15-4d49-b6c1-8f6418b2edc7/pull/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.489668 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-85txs_3af1207d-dab1-4b06-9dc4-0cfed43f145f/registry-server/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.500291 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-vdm72_94dce789-2750-40a1-b622-11ad58a438df/marketplace-operator/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.603571 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-utilities/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.781540 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-content/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.819560 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-utilities/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.838189 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-content/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.959839 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-content/0.log" Oct 02 12:24:36 crc kubenswrapper[4751]: I1002 12:24:36.972475 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/extract-utilities/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.188619 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-utilities/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.207688 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bxscj_4633ba02-9f3d-4de1-8214-0e908d7d3b72/registry-server/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.373866 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-content/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.412237 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-utilities/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.412351 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-content/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.567114 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-utilities/0.log" Oct 02 12:24:37 crc kubenswrapper[4751]: I1002 12:24:37.591909 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/extract-content/0.log" Oct 02 12:24:38 crc kubenswrapper[4751]: I1002 12:24:38.312697 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cqd2n_9b28c6e5-2f16-4e4c-8491-f0c685d33817/registry-server/0.log" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.387985 4751 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:39 crc kubenswrapper[4751]: E1002 12:24:39.389013 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="extract-content" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.389034 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="extract-content" Oct 02 12:24:39 crc kubenswrapper[4751]: E1002 12:24:39.389096 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="extract-utilities" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.389107 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="extract-utilities" Oct 02 12:24:39 crc kubenswrapper[4751]: E1002 12:24:39.389120 4751 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="registry-server" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.389127 4751 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="registry-server" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.389429 4751 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9b0cde-2aba-46a6-916e-753f83f22f02" containerName="registry-server" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.391026 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.404824 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.437986 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.438194 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xqlv\" (UniqueName: \"kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.438323 4751 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.540669 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.541310 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.541521 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xqlv\" (UniqueName: \"kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.541842 4751 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.542383 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.570393 4751 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xqlv\" (UniqueName: \"kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv\") pod \"community-operators-p2g4b\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:39 crc kubenswrapper[4751]: I1002 12:24:39.713667 4751 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:40 crc kubenswrapper[4751]: I1002 12:24:40.239328 4751 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:41 crc kubenswrapper[4751]: I1002 12:24:41.202825 4751 generic.go:334] "Generic (PLEG): container finished" podID="e7b7e96b-ffe3-4aff-86d3-712c780ffc23" containerID="64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa" exitCode=0 Oct 02 12:24:41 crc kubenswrapper[4751]: I1002 12:24:41.202870 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerDied","Data":"64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa"} Oct 02 12:24:41 crc kubenswrapper[4751]: I1002 12:24:41.202902 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerStarted","Data":"36a0ae0c128e2d01cf500aa345c60427d30d5b4bce38fbad0f3570dbb72209ce"} Oct 02 12:24:43 crc kubenswrapper[4751]: I1002 12:24:43.222709 4751 generic.go:334] "Generic (PLEG): container finished" podID="e7b7e96b-ffe3-4aff-86d3-712c780ffc23" containerID="051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164" exitCode=0 Oct 02 12:24:43 crc kubenswrapper[4751]: I1002 12:24:43.222821 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerDied","Data":"051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164"} Oct 02 12:24:45 crc kubenswrapper[4751]: I1002 12:24:45.242344 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerStarted","Data":"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12"} Oct 02 12:24:45 crc kubenswrapper[4751]: I1002 12:24:45.269359 4751 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p2g4b" podStartSLOduration=3.410221441 podStartE2EDuration="6.269341628s" podCreationTimestamp="2025-10-02 12:24:39 +0000 UTC" firstStartedPulling="2025-10-02 12:24:41.205013969 +0000 UTC m=+5563.259240419" lastFinishedPulling="2025-10-02 12:24:44.064134156 +0000 UTC m=+5566.118360606" observedRunningTime="2025-10-02 12:24:45.262612586 +0000 UTC m=+5567.316839036" watchObservedRunningTime="2025-10-02 12:24:45.269341628 +0000 UTC m=+5567.323568078" Oct 02 12:24:49 crc kubenswrapper[4751]: I1002 12:24:49.715454 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:49 crc kubenswrapper[4751]: I1002 12:24:49.715906 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:49 crc kubenswrapper[4751]: I1002 12:24:49.771286 4751 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:50 crc kubenswrapper[4751]: I1002 12:24:50.350432 4751 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:50 crc kubenswrapper[4751]: I1002 12:24:50.402890 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:52 crc kubenswrapper[4751]: I1002 12:24:52.323518 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p2g4b" podUID="e7b7e96b-ffe3-4aff-86d3-712c780ffc23" containerName="registry-server" containerID="cri-o://1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12" gracePeriod=2 Oct 02 12:24:52 crc kubenswrapper[4751]: I1002 12:24:52.921530 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.079657 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities\") pod \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.079759 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xqlv\" (UniqueName: \"kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv\") pod \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.079806 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content\") pod \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\" (UID: \"e7b7e96b-ffe3-4aff-86d3-712c780ffc23\") " Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.081804 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities" (OuterVolumeSpecName: "utilities") pod "e7b7e96b-ffe3-4aff-86d3-712c780ffc23" (UID: "e7b7e96b-ffe3-4aff-86d3-712c780ffc23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.106294 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv" (OuterVolumeSpecName: "kube-api-access-8xqlv") pod "e7b7e96b-ffe3-4aff-86d3-712c780ffc23" (UID: "e7b7e96b-ffe3-4aff-86d3-712c780ffc23"). InnerVolumeSpecName "kube-api-access-8xqlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.173489 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7b7e96b-ffe3-4aff-86d3-712c780ffc23" (UID: "e7b7e96b-ffe3-4aff-86d3-712c780ffc23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.182605 4751 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.182639 4751 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.182650 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xqlv\" (UniqueName: \"kubernetes.io/projected/e7b7e96b-ffe3-4aff-86d3-712c780ffc23-kube-api-access-8xqlv\") on node \"crc\" DevicePath \"\"" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.338541 4751 generic.go:334] "Generic (PLEG): container finished" podID="e7b7e96b-ffe3-4aff-86d3-712c780ffc23" containerID="1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12" exitCode=0 Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.338584 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerDied","Data":"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12"} Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.338610 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p2g4b" event={"ID":"e7b7e96b-ffe3-4aff-86d3-712c780ffc23","Type":"ContainerDied","Data":"36a0ae0c128e2d01cf500aa345c60427d30d5b4bce38fbad0f3570dbb72209ce"} Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.338612 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p2g4b" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.338627 4751 scope.go:117] "RemoveContainer" containerID="1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.385352 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.394742 4751 scope.go:117] "RemoveContainer" containerID="051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.402532 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p2g4b"] Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.415265 4751 scope.go:117] "RemoveContainer" containerID="64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.458949 4751 scope.go:117] "RemoveContainer" containerID="1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12" Oct 02 12:24:53 crc kubenswrapper[4751]: E1002 12:24:53.468920 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12\": container with ID starting with 1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12 not found: ID does not exist" containerID="1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.468988 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12"} err="failed to get container status \"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12\": rpc error: code = NotFound desc = could not find container \"1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12\": container with ID starting with 1d3a2642d75b9f3b29c0623210e713322e0d76e14df52462be3c676798f5ff12 not found: ID does not exist" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.469026 4751 scope.go:117] "RemoveContainer" containerID="051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164" Oct 02 12:24:53 crc kubenswrapper[4751]: E1002 12:24:53.469619 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164\": container with ID starting with 051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164 not found: ID does not exist" containerID="051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.469659 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164"} err="failed to get container status \"051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164\": rpc error: code = NotFound desc = could not find container \"051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164\": container with ID starting with 051f89fc9b9110ebcbd5d6739c46451445294856488d4f9d46d2edfd72ebc164 not found: ID does not exist" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.469682 4751 scope.go:117] "RemoveContainer" containerID="64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa" Oct 02 12:24:53 crc kubenswrapper[4751]: E1002 12:24:53.470119 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa\": container with ID starting with 64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa not found: ID does not exist" containerID="64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.470142 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa"} err="failed to get container status \"64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa\": rpc error: code = NotFound desc = could not find container \"64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa\": container with ID starting with 64f4da48ead384d3b01331e514b59eb902873c34caab0887858617bc401787fa not found: ID does not exist" Oct 02 12:24:53 crc kubenswrapper[4751]: I1002 12:24:53.559491 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b7e96b-ffe3-4aff-86d3-712c780ffc23" path="/var/lib/kubelet/pods/e7b7e96b-ffe3-4aff-86d3-712c780ffc23/volumes" Oct 02 12:24:54 crc kubenswrapper[4751]: E1002 12:24:54.690874 4751 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.163:43208->38.129.56.163:40831: write tcp 38.129.56.163:43208->38.129.56.163:40831: write: broken pipe Oct 02 12:24:57 crc kubenswrapper[4751]: E1002 12:24:57.049642 4751 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.163:43426->38.129.56.163:40831: read tcp 38.129.56.163:43426->38.129.56.163:40831: read: connection reset by peer Oct 02 12:25:01 crc kubenswrapper[4751]: I1002 12:25:01.507370 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:25:01 crc kubenswrapper[4751]: I1002 12:25:01.507906 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:25:01 crc kubenswrapper[4751]: I1002 12:25:01.507951 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:25:01 crc kubenswrapper[4751]: I1002 12:25:01.508791 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:25:01 crc kubenswrapper[4751]: I1002 12:25:01.508846 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec" gracePeriod=600 Oct 02 12:25:02 crc kubenswrapper[4751]: I1002 12:25:02.423509 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec" exitCode=0 Oct 02 12:25:02 crc kubenswrapper[4751]: I1002 12:25:02.423553 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec"} Oct 02 12:25:02 crc kubenswrapper[4751]: I1002 12:25:02.423969 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerStarted","Data":"daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8"} Oct 02 12:25:02 crc kubenswrapper[4751]: I1002 12:25:02.423991 4751 scope.go:117] "RemoveContainer" containerID="2055bab8e0ec0c488b0388d8dc13489238babbf107f14943368ebce18455efd4" Oct 02 12:26:34 crc kubenswrapper[4751]: I1002 12:26:34.402604 4751 generic.go:334] "Generic (PLEG): container finished" podID="2dc32e3a-c701-45f4-bdad-74dab7cfa711" containerID="9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92" exitCode=0 Oct 02 12:26:34 crc kubenswrapper[4751]: I1002 12:26:34.402668 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" event={"ID":"2dc32e3a-c701-45f4-bdad-74dab7cfa711","Type":"ContainerDied","Data":"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92"} Oct 02 12:26:34 crc kubenswrapper[4751]: I1002 12:26:34.403853 4751 scope.go:117] "RemoveContainer" containerID="9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92" Oct 02 12:26:34 crc kubenswrapper[4751]: I1002 12:26:34.481491 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5wn2f_must-gather-dqrtt_2dc32e3a-c701-45f4-bdad-74dab7cfa711/gather/0.log" Oct 02 12:26:42 crc kubenswrapper[4751]: I1002 12:26:42.739103 4751 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5wn2f/must-gather-dqrtt"] Oct 02 12:26:42 crc kubenswrapper[4751]: I1002 12:26:42.740335 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" podUID="2dc32e3a-c701-45f4-bdad-74dab7cfa711" containerName="copy" containerID="cri-o://aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8" gracePeriod=2 Oct 02 12:26:42 crc kubenswrapper[4751]: I1002 12:26:42.772147 4751 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5wn2f/must-gather-dqrtt"] Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.265954 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5wn2f_must-gather-dqrtt_2dc32e3a-c701-45f4-bdad-74dab7cfa711/copy/0.log" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.266582 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.301880 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output\") pod \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.302411 4751 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqvs4\" (UniqueName: \"kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4\") pod \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\" (UID: \"2dc32e3a-c701-45f4-bdad-74dab7cfa711\") " Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.308413 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4" (OuterVolumeSpecName: "kube-api-access-rqvs4") pod "2dc32e3a-c701-45f4-bdad-74dab7cfa711" (UID: "2dc32e3a-c701-45f4-bdad-74dab7cfa711"). InnerVolumeSpecName "kube-api-access-rqvs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.403988 4751 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqvs4\" (UniqueName: \"kubernetes.io/projected/2dc32e3a-c701-45f4-bdad-74dab7cfa711-kube-api-access-rqvs4\") on node \"crc\" DevicePath \"\"" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.480836 4751 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2dc32e3a-c701-45f4-bdad-74dab7cfa711" (UID: "2dc32e3a-c701-45f4-bdad-74dab7cfa711"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.504669 4751 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2dc32e3a-c701-45f4-bdad-74dab7cfa711-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.505366 4751 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5wn2f_must-gather-dqrtt_2dc32e3a-c701-45f4-bdad-74dab7cfa711/copy/0.log" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.505853 4751 generic.go:334] "Generic (PLEG): container finished" podID="2dc32e3a-c701-45f4-bdad-74dab7cfa711" containerID="aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8" exitCode=143 Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.505906 4751 scope.go:117] "RemoveContainer" containerID="aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.505967 4751 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5wn2f/must-gather-dqrtt" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.529970 4751 scope.go:117] "RemoveContainer" containerID="9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.563191 4751 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc32e3a-c701-45f4-bdad-74dab7cfa711" path="/var/lib/kubelet/pods/2dc32e3a-c701-45f4-bdad-74dab7cfa711/volumes" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.607848 4751 scope.go:117] "RemoveContainer" containerID="aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8" Oct 02 12:26:43 crc kubenswrapper[4751]: E1002 12:26:43.608264 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8\": container with ID starting with aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8 not found: ID does not exist" containerID="aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.608295 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8"} err="failed to get container status \"aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8\": rpc error: code = NotFound desc = could not find container \"aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8\": container with ID starting with aef26df5c06bee28684c6845656221864d9751236230ab5a73a711633438a7f8 not found: ID does not exist" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.608319 4751 scope.go:117] "RemoveContainer" containerID="9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92" Oct 02 12:26:43 crc kubenswrapper[4751]: E1002 12:26:43.608820 4751 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92\": container with ID starting with 9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92 not found: ID does not exist" containerID="9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92" Oct 02 12:26:43 crc kubenswrapper[4751]: I1002 12:26:43.608848 4751 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92"} err="failed to get container status \"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92\": rpc error: code = NotFound desc = could not find container \"9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92\": container with ID starting with 9c2c5c244cfe4acb0a3406664b44c2b24733c5c68fb9d89511b78da766b29b92 not found: ID does not exist" Oct 02 12:27:01 crc kubenswrapper[4751]: I1002 12:27:01.507085 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:27:01 crc kubenswrapper[4751]: I1002 12:27:01.507716 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:27:13 crc kubenswrapper[4751]: I1002 12:27:13.251942 4751 scope.go:117] "RemoveContainer" containerID="eea132963f1088755b5e8f2a9a6f3b675eee4f688c933a1f5977470960a8e1e6" Oct 02 12:27:13 crc kubenswrapper[4751]: I1002 12:27:13.278605 4751 scope.go:117] "RemoveContainer" containerID="cd8e0c7f3c56cc48b152a71ea571312aefda758ede2149c7a846a81282b7b4b1" Oct 02 12:27:13 crc kubenswrapper[4751]: I1002 12:27:13.317812 4751 scope.go:117] "RemoveContainer" containerID="d54b082a9ea5a157cfdbfcc7541c1ccb29ae9eb095ae1d6f90ca399acbc369a7" Oct 02 12:27:31 crc kubenswrapper[4751]: I1002 12:27:31.507093 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:27:31 crc kubenswrapper[4751]: I1002 12:27:31.508216 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:28:01 crc kubenswrapper[4751]: I1002 12:28:01.507661 4751 patch_prober.go:28] interesting pod/machine-config-daemon-b4wsd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 12:28:01 crc kubenswrapper[4751]: I1002 12:28:01.508349 4751 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 12:28:01 crc kubenswrapper[4751]: I1002 12:28:01.508423 4751 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" Oct 02 12:28:01 crc kubenswrapper[4751]: I1002 12:28:01.509266 4751 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8"} pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 12:28:01 crc kubenswrapper[4751]: I1002 12:28:01.509338 4751 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerName="machine-config-daemon" containerID="cri-o://daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" gracePeriod=600 Oct 02 12:28:02 crc kubenswrapper[4751]: E1002 12:28:02.258983 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:28:02 crc kubenswrapper[4751]: I1002 12:28:02.282392 4751 generic.go:334] "Generic (PLEG): container finished" podID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" containerID="daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" exitCode=0 Oct 02 12:28:02 crc kubenswrapper[4751]: I1002 12:28:02.282460 4751 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" event={"ID":"173e3cf6-4beb-4df2-a3f8-80504ed406cc","Type":"ContainerDied","Data":"daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8"} Oct 02 12:28:02 crc kubenswrapper[4751]: I1002 12:28:02.282522 4751 scope.go:117] "RemoveContainer" containerID="533b56745a69619114fc5965906bbee8a8953c088a5f857d4b39f0c82f0ac2ec" Oct 02 12:28:02 crc kubenswrapper[4751]: I1002 12:28:02.283134 4751 scope.go:117] "RemoveContainer" containerID="daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" Oct 02 12:28:02 crc kubenswrapper[4751]: E1002 12:28:02.283409 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:28:15 crc kubenswrapper[4751]: I1002 12:28:15.550748 4751 scope.go:117] "RemoveContainer" containerID="daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" Oct 02 12:28:15 crc kubenswrapper[4751]: E1002 12:28:15.551911 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:28:26 crc kubenswrapper[4751]: I1002 12:28:26.550073 4751 scope.go:117] "RemoveContainer" containerID="daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" Oct 02 12:28:26 crc kubenswrapper[4751]: E1002 12:28:26.550813 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" Oct 02 12:28:39 crc kubenswrapper[4751]: I1002 12:28:39.560150 4751 scope.go:117] "RemoveContainer" containerID="daaeab03aa2d0ee96a756d42b81ee9ae74f06f8c99044f7a67eac82f974f12b8" Oct 02 12:28:39 crc kubenswrapper[4751]: E1002 12:28:39.560943 4751 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b4wsd_openshift-machine-config-operator(173e3cf6-4beb-4df2-a3f8-80504ed406cc)\"" pod="openshift-machine-config-operator/machine-config-daemon-b4wsd" podUID="173e3cf6-4beb-4df2-a3f8-80504ed406cc" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067470007024453 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067470007017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067454052016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067454053015466 5ustar corecore